TLDR: A novel hardware-based attack allows adversaries to steal AI model weights from wireless devices equipped with AI accelerators. This is achieved by implanting a stealthy hardware Trojan (HT) that creates a hidden communication channel within legitimate wireless transmissions. The HT subtly leaks model parameters, which are then intercepted and reconstructed by a nearby adversary. The attack is agnostic to AI model architecture and hardware, validated across diverse models, and can reliably exfiltrate even large models within hours, using error mitigation techniques like repetition and voting to maintain accuracy. Detecting this attack is challenging due to the HT’s minimal footprint and the covert channel’s transparency.
Artificial intelligence models are incredibly valuable assets, representing significant investments in data, computing power, and training time. They also provide a competitive edge and often embody proprietary techniques. This makes them prime targets for theft, a serious concern for AI model providers.
A new and sophisticated attack has been uncovered that targets wireless devices equipped with AI hardware accelerators. This novel method allows attackers to covertly steal AI model weights, the core components of an AI model, without the victim ever realizing it. The attack is designed to be highly stealthy and effective, regardless of the specific AI model architecture or the hardware accelerator being used.
How the Attack Unfolds
The attack operates in two distinct phases. In the first phase, the victim’s device is compromised with a ‘hardware Trojan’ (HT). This isn’t a software virus, but a malicious modification embedded directly into the device’s hardware. This HT is specifically designed to create a hidden communication channel, allowing it to secretly leak model weights.
The second phase involves the adversary, who uses a nearby wireless device to intercept the victim’s transmission frames during normal operation. As the victim’s device communicates, the HT subtly embeds fragments of the AI model’s weight matrix into these legitimate transmissions. The adversary then incrementally reconstructs the complete weight matrix from these intercepted fragments.
The Stealthy Hardware Trojan and Covert Channel
Unlike traditional software attacks that might query a model through an API, this is a hardware-based attack. It doesn’t require physical access to the device during the stealing phase; the attacker only needs to be within wireless communication range. The hardware Trojan is designed with a minimal footprint, making it incredibly difficult to detect through conventional methods like reverse engineering or logic testing. It causes negligible overhead in terms of logic utilization or power consumption, further enhancing its stealth.
The covert channel itself is hidden within the preamble of standard wireless communication frames, specifically in the Short Training Sequence (STS) of Wi-Fi transmissions. This part of the transmission is normally used for synchronization and is typically discarded by regular receivers. However, the HT subtly modulates this STS with bits of the AI model’s weights. The researchers found that by carefully selecting a parameter called ‘alpha’ (specifically, 15%), the covert channel becomes imperceptible to a normal receiver, meaning it doesn’t degrade communication performance or raise suspicion.
Who is the Adversary?
The threat model for this attack suggests that the adversary could be the design house of the device, the foundry that manufactures it, or even a third-party attacker who commissions these entities to facilitate the attack. The victim, unaware of the compromise, loads their valuable AI model onto the device for inference, and the HT begins its secret exfiltration during routine wireless operations.
Reconstructing the Stolen Model
The adversary, referred to as Eve, uses a specialized receiver to intercept and process the modified STS, extracting the hidden bits. Because wireless communication is prone to errors (Bit Error Rate or BER), Eve might initially reconstruct an approximate weight matrix. However, AI models can be surprisingly resilient to some bit flips. For higher error rates, Eve can employ a clever error mitigation technique: receiving multiple broadcasts of the weight matrix and applying a ‘voting scheme’. By comparing multiple copies of each bit, the most frequent value is selected, effectively reducing the BER and restoring the model’s accuracy to its baseline level.
Demonstrated Effectiveness
The researchers validated their approach through a hardware-based demonstration using two bladeRF 2.0 micro xA9 boards. They tested the attack against four diverse AI models: LeNet-5 (image classification), MobileNetV3-Large (image classification), IBM DVS128 Gesture SNN (Spiking Neural Network), and YOLOv11n (object detection). The leakage time varied from a few seconds for smaller models like LeNet-5 to approximately two hours for the largest model, MobileNetV3-Large, under favorable conditions.
The study showed that larger models tend to be more sensitive to BER, meaning their accuracy drops at lower error rates. Interestingly, quantized versions of models (using 8-bit integers instead of 32-bit floating-point) were found to be more robust to bit flips. Under high Signal-to-Noise Ratio (SNR) conditions, models could be leaked in a single broadcast with baseline accuracy. In more challenging, lower SNR environments, a few repetitions of the leakage process combined with the voting scheme were sufficient to restore accuracy.
Also Read:
- Adversarial Attacks Threaten Deep Learning Security in SDN-IoT Networks
- The Stealthy Threat of Deceptive AI Reasoning: Introducing DecepChain
Challenges in Defense
Detecting such a sophisticated hardware Trojan and covert channel is extremely difficult for a victim who doesn’t have access to the original design files. Traditional methods like reverse engineering are costly and time-consuming, and the HT’s minimal footprint makes it hard to spot. Logic testing and statistical side-channel fingerprinting often require a ‘golden reference’ chip or design models, which the victim lacks. Even advanced AI-based detection mechanisms, while promising, might fail against novel covert channel strategies not represented in their training data.
This research highlights a significant new threat in AI security, demonstrating how valuable AI models can be stolen from edge devices through highly stealthy hardware modifications and covert wireless communication. For more technical details, you can read the full paper here.


