Newsletter →
HackerDose
HackerDose
  • Latest Stories
  • Security & Tech
    • Cybersecurity
    • Technology
    • Vulnerabilities
    • Dark Web
  • Crypto & Blockchain
    • Cryptocurrency
    • Blockchain
    • Finance
    • Price Predictions
    • Guides
    • Regulation
Reading: Sleeping Backdoor in AI Models Waiting to Strike
Newsletter
Newsletter →
HackerDose
HackerDose
  • Latest Stories
  • Security & Tech
    • Cybersecurity
    • Technology
    • Vulnerabilities
    • Dark Web
  • Crypto & Blockchain
    • Cryptocurrency
    • Blockchain
    • Finance
    • Price Predictions
    • Guides
    • Regulation
Reading: Sleeping Backdoor in AI Models Waiting to Strike
Newsletter
Search
  • Latest Stories
  • Security & Tech
    • Security
    • Vulnerabilities
    • Dark Web
    • Technology
    • Privacy
  • Crypto & Blockchain
    • Cryptocurrency
    • Blockchain
    • Finance
    • Price Predictions
    • Guides
    • Regulation
© MRS Media Company. Hackerdose LLC. All Rights Reserved.

Security » Sleeping Backdoor in AI Models Waiting to Strike

SecurityVulnerabilities

Sleeping Backdoor in AI Models Waiting to Strike

Hackers can now hijack neural networks undetected using ShadowLogic, a new backdoor technique.

Marco Rizal
Last updated: October 18, 2024 12:20 pm
By Marco Rizal - Editor, Journalist 3 Min Read
Share
Backdoor in AI Models Waiting to Strike
SHARE
  • ShadowLogic allows attackers to implant backdoors without modifying code.
  • Backdoors persist even through model fine-tuning.
  • This attack can target any AI model, making it a supply chain risk

HiddenLayer's Security, Artificial Intelligence (SAI) team has discovered a new method for backdooring neural networks.

Dubbed as “ShadowLogic,” this technique enables attackers to insert backdoors into any neural network model by manipulating its computational graph, bypassing traditional code-based exploits.

This approach has major consequences for AI models because it introduces a new way for adversaries to hijack systems without changing any weights or biases.

ShadowLogic works within a model's architecture, embedding a backdoor into the computational graph—the framework that governs how neural networks process data.

Unlike previous attacks, this method can withstand fine-tuning, allowing the backdoor to remain operational even as models are updated.

When the hidden “logic” is triggered by specific inputs, the model produces attacker-defined results, transforming trusted AI applications into dangerous tools.

HiddenLayer warns that this vulnerability endangers the AI supply chain, making every fine-tuned model a security risk.

Neural network models, particularly large foundation models, are ideal candidates for this technique.

They are widely used across industries and frequently repurposed in downstream applications ranging from image classification to fraud detection.

If the attack is carried out in critical sectors where AI is used to make decisions, it could have disastrous consequences.

The researchers demonstrated how a bad actor could hijack AI models without leaving a trace.

It's not just about training phase backdoors anymore; now, an attacker can modify how a model computes without touching its training data or weights.

ShadowLogic is not merely conceptual. It has been successfully tested on well-known architectures such as ResNet, which is widely used for image classification.

The researchers used a simple visual trigger, a red square in an image to show how the backdoor could manipulate the model's output.

The model's predictions changed when the red square appeared, despite the fact that the original image had no visible corruption.

image 26
Original vs triggered images with red trigger (via: HiddenLayer)

Worse still, these triggers do not have to be obvious. The visual marker used in demonstrations could be rendered imperceptible, allowing an adversary to subtly modify an image to cause malicious behavior in real-world applications.

The HiddenLayer team also broadened their experiments to include other architectures, such as the YOLO model for object detection and the Phi-3 small language models.

These trials were equally successful, demonstrating that models other than image classifiers are vulnerable.

For example, by defining a specific input phrase, researchers were able to control the Phi-3 model's output, replacing legitimate responses with attacker-defined content.

ShadowLogic is not the first attack technique to target AI models. Previous research from New York University and UC Berkeley investigated backdoors inserted during the training phase.

But HiddenLayer's technique goes a step further, demonstrating that backdoors can be embedded without the need for complex retraining.

Leave a comment Leave a comment

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Subscribe to our newsletter

Receive the latest news and stories straight to your inbox.

Latest stories

Bitcoin Holds at $85K as Global Trade Tensions and Fed Speculation Unfold

April 15, 2025

Michael Saylor Doubles Down on Bitcoin (BTC) with $285M Investment Amid Global Uncertainty

April 14, 2025

Mantra Faces Crisis After OM Token Crashes 90% in a Day

April 14, 2025

Solana (SOL) on the Verge of a Breakout: Could $300 Be the Next Target?

April 14, 2025

You might also like

Two Thirds of The Internet Is a Bot Playground

Two-Thirds of The Internet Is a Bot Playground

Bitcoins DoS Vulnerability

Bitcoin’s DoS Vulnerability Lets Hackers Crash Miners For Less Than 1% of a Block

Hackers Are Buying Big YouTube Channels to Infect You with Malware

Hackers Are Buying Big YouTube Channels to Infect You with Malware

Perfctl Malware Exploits Dockers Wide Open Remote API Servers

Perfctl Malware Exploits Docker’s Wide-Open Remote API Servers

Newsletter

Our website stores cookies on your computer. They allow us to remember you and help personalize your experience with our site

Quick Links

  • Contact Us
  • Search
  • Malware
  • Downloads

Company

  • About Us
  • Terms and Conditions
  • Cookies Policy
  • Privacy Policy
Advertise with us

Socials

Follow Us

© 2025 | HackerDose Media Company – All Rights Reserved

Welcome Back!

Sign in to your account

Lost your password?