Adversarial Attacks: Hacking AI in Surprising Ways

When AI sees what isn't there

Adversarial attack concept

I still remember the first time I saw an adversarial attack. I showed an image to a classifier—it confidently said "panda." Then someone added barely-perceptible noise, showed it again, and the model confidently said "gibbon." The image still looked like a panda to me. That's when I realized AI sees the world very differently than humans.

What Are Adversarial Examples?

Adversarial examples are inputs that have been specially crafted to cause AI models to make mistakes. The changes are often tiny—sometimes invisible to the human eye—but can completely fool the model.

It's not just images. Adversarial attacks work on text, audio, and any domain where neural networks operate.

How Do They Work?

The key insight: neural networks learn statistical correlations, not true understanding. They find patterns that work most of the time—but those patterns can be exploited.

In image classification, small changes to pixels can shift the prediction. The model isn't "seeing" the image the way we do—it's responding to mathematical features that happen to correlate with the label.

Types of Attacks

White-box attacks: The attacker has full access to the model—architecture, weights, gradients. They can compute the exact modification needed to fool the model.

Black-box attacks: The attacker can only query the model and see outputs. They build a substitute model and attack that, then transfer to the real model.

Physical attacks: Instead of digital images, the adversarial perturbation is printed and placed in the real world. A sticker on a stop sign can make it invisible to self-driving cars.

Real-World Implications

This is serious. Consider:

Defenses

Adversarial training: Train on adversarial examples. Makes the model more robust but can hurt accuracy on clean data.

Input preprocessing: Denoising, compression, randomization can disrupt adversarial perturbations.

Detection: Train a separate model to detect adversarial inputs.

Certified robustness: Mathematical guarantees on worst-case performance.

The arms race continues. As we deploy AI in safety-critical applications, adversarial robustness becomes essential.