Security

Research

Boffins trick AI model into giving up its secrets

All it took to make an Google Edge TPU give up model hyperparameters was specific hardware, a novel attack technique … and several days


Computer scientists from North Carolina State University have devised a way to copy AI models running on Google Edge Tensor Processing Units (TPUs), as used in Google Pixel phones and third-party machine learning accelerators.

The technique, developed by NC State researchers Ashley Kurian, Anuj Dubey, Ferhat Yaman and Aydin Aysu, is a side-channel attack that measures the electromagnetic intensity of AI model use (inference) when running on TPUs, and exploits those measurements to infer model hyperparameters.

Machine learning model hyperparameters refer to values set prior to the training process that affect model training – the learning rate, the batch size, or the pool size. They're distinct from model parameters – such as weights – which are internal to the model and are learned during training.

An adversary with both can mostly reproduce an AI model at far less cost than incurred during the original training process – something developers spending billions on building AI models might prefer to avoid. There are already a variety of parameter extraction techniques.

"A hyperparameter stealing attack followed by parameter extraction can create a high-fidelity substitute model with the extracted information to mimic the victim model," the researchers explain in their paper, "TPUXtract: An Exhaustive Hyperparameter Extraction Framework."

While there have been prior limited hyperparameter attacks, the researchers claim their attack is the first to perform a comprehensive hyperparameter extraction and the first model stealing attack targeting the Google Edge TPU.

"Because we stole the architecture and layer details, we were able to recreate the high-level features of the AI," explained Aydin Aysu, a co-author of the paper and associate professor at NC State, in a statement. "We then used that information to recreate the functional AI model, or a very close surrogate of that model."

The attack scenario assumes the adversary has access to the device – a Coral Dev Board with a Google Edge TPU – during inference, and can conduct electromagnetic measurements using Riscure hardware (icWaves, Transceiver, High Sensitivity EM probe) and a PicoScope Oscilloscope. Knowledge of the software deployment environment (TF Lite for Edge TPU) is also assumed. However, the details about Edge TPU's architecture and instruction set are not required.

The researchers' approach involves extracting information about each neural network layer sequentially and then feeding extracted hyperparameters for each layer back into the layer extraction framework. This overcomes problems with prior efforts that required an impractical brute force attack against the entire model but yielded only some of the model's hyperparameters.

According to the researchers, their approach is able to recreate a model with 99.91 percent accuracy. The process – tested on models such as MobileNet V3, Inception V3, and ResNet-50 – takes about three hours per layer. The models cited in the paper range from 28 to 242 layers.

"Our research demonstrates that an adversary can effectively reverse engineer the hyperparameters of a neural network by observing its EM emanations during inference, even in a black box setting," the authors state in their paper. "The coverage and accuracy of our approach raise significant concerns about the vulnerability of commercial accelerators like the Edge TPU to model stealing in various real-world scenarios."

Google is aware of the researchers' findings, and declined to comment on the record. The Register understands from conversations with shy comms folk that one of the reasons the Coral Dev Board was chosen is that it does not implement memory encryption. ®

Send us news
20 Comments

Google reports halving code migration time with AI help

Chocolate Factory slurps own dogfood, sheds drudgery in specific areas

Microsoft eggheads say AI can never be made secure – after testing Redmond's own products

If you want a picture of the future, imagine your infosec team stamping on software forever

Biden signs sweeping cybersecurity order, just in time for Trump to gut it

Ransomware, AI, secure software, digital IDs – there's something for everyone in the presidential directive

Sage Copilot grounded briefly to fix AI misbehavior

'Minor issue' with showing accounting customers 'unrelated business information' required repairs

OpenAI's ChatGPT crawler can be tricked into DDoSing sites, answering your queries

The S in LLM stands for Security

Just as your LLM once again goes off the rails, Cisco, Nvidia are at the door smiling

Some of you have apparently already botched chatbots or allowed ‘shadow AI’ to creep in

Megan, AI recruiting agent, is on the job, giving bosses fewer reasons to hire in HR

She doesn't feel pity, remorse, or fear, but she'll craft a polite email message as she turns you down

3Blue1Brown copyright takedown blunder by AI biz blamed on human error

Worker copy-pasted wrong YouTube URL, says ChainPatrol

Google and Linux Foundation form Chromium love club

Right as Uncle Sam pushes for Chrome sell-off, eh?

Free-software warriors celebrate landmark case that enforced GNU LGPL

On the Fritz: German router maker AVM lets device rights case end after coughing up source code

Look for the label: White House rolls out 'Cyber Trust Mark' for smart devices

Beware the IoT that doesn’t get a security tag

Uncle Sam now targets six landlord giants in war on alleged algorithmic rent fixing

One of ya is gonna sing like a canary, prosecutors say