ANALYTICAL BRIEFREF: ADVR-0326-AI|SOURCE: OSINT / NIST / WEST POINT LIEBER INSTITUTE / ACADEMIC LITERATURE
UPDATED 16 MAR 2026
THE POISONED WELL

THE CORRUPTED MODEL

Data Poisoning, Adversarial ML, and the Silent Attack Against Every AI System on the Battlefield

SUBJECT Adversarial Machine Learning in Military AI
REGION Global
PRIORITY HIGH
ANALYST OPEN SOURCE
STATUS ACTIVE THREAT
NIST AI 100-2: formal taxonomy of adversarial ML attacks — poisoning, evasion, model extraction, privacy attacks ///West Point Lieber Institute: "By covertly introducing manipulated data, adversarial AI systems can be rendered ineffective" ///Gartner: 30% of all AI cyberattacks will leverage training-data poisoning or adversarial samples by 2028 ///UC Irvine FlyTrap already proved: a $20 printed pattern defeats neural network object tracking in the physical world ///Maven Smart System generated 1,000 targets in 24 hours using AI — but what if the training data was compromised? ///Nation-states are shifting from hacking networks to poisoning the AI models those networks depend on ///NIST AI 100-2: formal taxonomy of adversarial ML attacks — poisoning, evasion, model extraction, privacy attacks ///West Point Lieber Institute: "By covertly introducing manipulated data, adversarial AI systems can be rendered ineffective" ///Gartner: 30% of all AI cyberattacks will leverage training-data poisoning or adversarial samples by 2028 ///UC Irvine FlyTrap already proved: a $20 printed pattern defeats neural network object tracking in the physical world ///Maven Smart System generated 1,000 targets in 24 hours using AI — but what if the training data was compromised? ///Nation-states are shifting from hacking networks to poisoning the AI models those networks depend on ///

THE AI YOUR AI CANNOT TRUST

GAITHERSBURG, MD — JANUARY 2025 | NIST AI 100-2

NIST Publishes Formal Taxonomy of Adversarial Machine Learning Attacks

In January 2025, the National Institute of Standards and Technology published NIST AI 100-2e2025 — the first formal U.S. government taxonomy of adversarial machine learning attacks.[1] The document categorizes the attack surface of AI systems into four classes: poisoning attacks (corrupting training data), evasion attacks (fooling deployed models), privacy attacks (extracting training data from models), and model extraction attacks (stealing the model architecture itself).[1]

The publication was not academic exercise. It was a formal acknowledgment that every AI system deployed by the U.S. government — including those running military targeting, intelligence analysis, and autonomous weapons — can be systematically corrupted by an adversary who understands how the model was trained. The attack surface is not the network. It is the data.

ATTACK CLASSES
4
Poisoning, evasion, privacy extraction, model theft — NIST AI 100-2[1]
PROJECTED SHARE
30%
Gartner: 30% of AI cyberattacks will use adversarial ML techniques by 2028[2]
COST TO ATTACK
$20
UC Irvine proved physical-world adversarial attack with a printed umbrella pattern[3]

By covertly introducing manipulated data during the training phase, adversarial AI systems can be rendered ineffective, misclassifying U.S. assets or misinterpreting battlefield conditions.

— West Point Lieber Institute, 2025[4]

DATA POISONING — CORRUPTION AT THE SOURCE

Data poisoning attacks inject malicious data into a model's training set before or during training. The corrupted model appears to function normally on standard inputs but behaves predictably wrong when it encounters specific trigger patterns — patterns chosen by the attacker.[1]

The military implications are catastrophic. Consider a targeting AI trained on satellite imagery to identify enemy vehicles. If an adversary poisons the training data — subtly mislabeling a fraction of images so that vehicles in a specific configuration are classified as civilian — the model will systematically miss those targets in the field. The AI performs perfectly in testing. It fails precisely when it matters.[4]

The West Point Lieber Institute analysis is blunt: poisoned military AI systems could "misclassify U.S. assets or misinterpret battlefield conditions" — causing friendly fire, missed threats, or strategic miscalculation. The attack is invisible to operators who trust the model's outputs because every standard validation metric shows the model is working correctly.[4]

Backdoor poisoning is the most dangerous variant. The attacker inserts a hidden "trigger" — a specific pixel pattern, metadata tag, or input feature — that activates malicious behavior only when present. In all other cases, the model performs normally. The backdoor survives retraining, fine-tuning, and standard security audits because it is encoded in the model's weights, not in the code.[1]

EVASION — FOOLING THE DEPLOYED MODEL

Evasion attacks target models already in deployment, crafting inputs that cause misclassification without modifying the model itself. UC Irvine's FlyTrap demonstrated this in the physical world: a printed pattern on an umbrella caused autonomous drone tracking AI to misidentify and physically chase the decoy, flying into a capture device.[3]

FlyTrap is the $20 proof of concept. But the principle scales. Adversarial patches applied to military vehicles can cause image-recognition AI to misclassify them — a tank becomes a truck, a mobile launcher becomes a civilian bus. Adversarial perturbations in radar returns can cause signal-processing AI to misidentify friend from foe. Adversarial inputs to natural language AI can cause intelligence analysis systems to draw incorrect conclusions from intercepted communications.[1]

The critical vulnerability: military AI systems — Maven's targeting engine, Pulsar's signal classification, autonomous drone navigation — all use the same fundamental neural network architectures that are susceptible to adversarial evasion. The mathematical vulnerability is inherent to how neural networks learn to classify. No currently deployed defense completely eliminates adversarial evasion attacks.[1]

Gartner projects that by 2028, 30% of all AI-related cyberattacks will leverage training-data poisoning, adversarial samples, or model theft — shifting the attack surface from networks and endpoints to the AI models themselves.[2] The battleground is no longer the firewall. It is the training pipeline.

THE OPEN-SOURCE TROJAN HORSE

Modern military AI is not built from scratch. It is built on top of open-source foundations: PyTorch, TensorFlow, Hugging Face model repositories, pre-trained weights from public datasets. Each dependency is a potential attack surface.

A supply chain poisoning attack targets the shared infrastructure of AI development: pre-trained models downloaded thousands of times, popular training datasets used as benchmarks, open-source libraries embedded in military AI pipelines. Compromising a widely-used pre-trained model or dataset can propagate poisoned weights into every downstream system that fine-tunes from that base.[1]

The parallel to the Shadow Brokers leak is direct. The Leak demonstrated that offensive cyber tools, once built, eventually proliferate beyond their creators' control. The Poisoned Well demonstrates the inverse: defensive AI tools, once deployed, can be corrupted from their foundations without anyone touching the deployed system. The attack travels through the supply chain, not the network.

The Pentagon's AI supply chain depends on commercial models and datasets. Claude runs inside Palantir's Maven Smart System. OpenAI models are being integrated into classified networks. The foundation models themselves are trained on internet-scale datasets that no human has fully audited. The question is not whether military AI training data has been compromised. It is whether anyone would know if it had been.

WHO IS DOING THIS

VentureBeat reported in December 2025 that nation-states are shifting from traditional network intrusions to adversarial ML attacks: "Disrupting entire networks with adversarial ML attacks is the stealth attack strategy nation-states are adopting."[2] The logic is straightforward: hacking a network leaves forensic traces. Poisoning a training dataset leaves no trace in the deployed system's code — only in its behavior, and only under specific conditions the attacker controls.

China has published extensively on adversarial ML in military contexts. PLA-affiliated researchers have authored papers on adversarial attacks against image recognition, radar signal processing, and autonomous navigation systems. This research is dual-use: understanding adversarial attacks is necessary for both offense (poisoning adversary AI) and defense (hardening your own).[5]

Russia's electronic warfare doctrine — the same doctrine that produced the Moscow Signal and the Havana Syndrome lineage — naturally extends to adversarial ML. If the electromagnetic spectrum is a contested domain, and AI systems are the primary consumers of spectrum data (radar, signals intelligence, communications), then attacking the AI's perception of the spectrum is the logical evolution of electronic warfare.

The convergence is clear: adversarial ML is the cyber-electromagnetic equivalent of poisoning a well. You don't attack the water. You attack the source. And everyone who drinks from it is compromised.

THE EMERGING THREAT

2013
Szegedy et al. publish the first systematic study of adversarial examples in neural networks — demonstrating that imperceptible perturbations to inputs cause confident misclassification. The fundamental vulnerability is identified.[6]
2017
Physical-world adversarial attacks demonstrated: printed patches cause image classifiers to misidentify stop signs and other objects. The threat moves from digital to physical.[7]
2018-2024
Military AI deployment accelerates: Project Maven, Palantir AIP, Anduril Lattice, Pulsar EW system. Each system uses neural networks susceptible to the same fundamental adversarial vulnerabilities.
JAN 2025
NIST publishes AI 100-2e2025: formal taxonomy of adversarial ML attacks. First official U.S. government framework for understanding the attack surface of deployed AI systems.[1]
MAR 2025
UC Irvine's FlyTrap paper: $20 adversarial pattern defeats autonomous drone tracking AI in the physical world. Proof that the theoretical vulnerability is practically exploitable against military-grade systems.[3]
JUN 2025
West Point Lieber Institute publishes analysis of data poisoning as a covert weapon against U.S. military AI superiority. Warns of misclassification of U.S. assets and misinterpretation of battlefield conditions.[4]
DEC 2025
VentureBeat: nation-states shifting to adversarial ML as primary attack vector. Gartner projects 30% of AI cyberattacks will use adversarial techniques by 2028.[2]

BOTTOM LINE

The United States has committed to AI-driven warfare. Operation Epic Fury generated 1,000 targets in 24 hours using Maven's AI. Pulsar learns to jam signals autonomously. Autonomous drones navigate without human control. Every one of these systems depends on neural networks that are mathematically vulnerable to adversarial manipulation.[1]

Data poisoning is the most dangerous vector because it is invisible. A poisoned model passes every standard test. It performs correctly on every benchmark. It fails only when the attacker's trigger is present — on the battlefield, at the moment of maximum consequence. West Point has warned that this could cause U.S. military AI to misclassify friendly assets or misread battlefield conditions.[4]

FlyTrap proved the principle with a $20 umbrella. The question is what a nation-state can do with the same mathematical insight, applied to the training data of models running inside classified networks. The AI kill chain is only as trustworthy as the data it was trained on — and no currently deployed audit can guarantee that data hasn't been compromised.

The Poisoned Well is the quiet counterpart to the visible arms race in autonomous weapons. While the Pentagon builds AI that fights, adversaries are learning to corrupt the AI before it reaches the battlefield. The next war may not be won by whoever has the best AI — but by whoever best understands the other side's training data.

Disrupting entire networks with adversarial ML attacks is the stealth attack strategy nation-states are adopting.

— VentureBeat, December 2025[2]

References & Source Material

  1. [1]NIST, "Adversarial Machine Learning: A Taxonomy and Terminology of Attacks and Mitigations," NIST AI 100-2e2025, January 2025. Formal taxonomy of poisoning, evasion, privacy, and extraction attacks.
  2. [2]VentureBeat, "Adversarial attacks on AI models are rising: what should you do now?" 22 Dec 2025. Gartner 30% projection, nation-state adversarial ML adoption.
  3. [3]UC Irvine, "FlyTrap: Adversarial Physical Countermeasures Against Autonomous Drone Tracking," 2025. $20 printed pattern defeats neural network object tracking.
  4. [4]West Point Lieber Institute, "Data Poisoning as a Covert Weapon: Securing U.S. Military Superiority in AI-Driven Warfare," 30 Jun 2025. Military AI poisoning threat analysis.
  5. [5]Various PLA-affiliated publications on adversarial machine learning in military contexts. Analyzed by Georgetown CSET.
  6. [6]Szegedy C. et al., "Intriguing properties of neural networks," ICLR 2014. First systematic study of adversarial examples.
  7. [7]Eykholt K. et al., "Robust Physical-World Attacks on Deep Learning Visual Classification," CVPR 2018. Physical-world adversarial attacks on road signs.
  8. [8]Palo Alto Networks, "What Are Adversarial AI Attacks on Machine Learning?" Overview of poisoning, evasion, and model theft attack vectors.
  9. [9]ISACA, "Combating the Threat of Adversarial Machine Learning to AI-Driven Cybersecurity," 2025. ML classifier evasion techniques.
CONNECTIONS
ZOOM OUT