AI Geopolitics · · 9 min read

The Human-in-the-Loop Mirage: How Combat Pressure Erased AI Warfare Safeguards

Pentagon blacklisting of Anthropic exposes systematic erosion of oversight protocols as Iran conflict validates speed gains—and civilian costs—of autonomous targeting at scale.

The Pentagon’s designation of Anthropic as a ‘supply chain risk’ on March 3, 2026—the first time the Cold War-era statute was applied to a domestic American company—marks the moment human oversight of military AI shifted from technical requirement to contractual inconvenience. What began as a contract dispute over usage restrictions has exposed a structural failure: when operational tempo demands decisions in seconds rather than minutes, the ‘human-in-the-loop’ safeguard collapses not from technical constraints but from deliberate policy choices that prioritise speed over accountability.

Iran Conflict: AI vs Human Performance
Maven AI Accuracy60%
Human Analyst Accuracy84%
Targets Processed (24hrs)1,000
Minab School Casualties165+

The Contractual Battlefield

The confrontation began in late February 2026 when Defense Secretary Pete Hegseth threatened Anthropic with a $200 million contract cancellation and Defense Production Act invocation unless the company agreed to ‘all lawful purposes’ usage—removing restrictions on Autonomous Weapons and domestic surveillance, according to AIThinkerLab. Anthropic had requested contractual language mirroring existing Pentagon policy—DOD Directive 3000.09, which requires ‘appropriate levels of human judgment’ for lethal force decisions. The Pentagon refused.

The company’s position hardened after January 2026, when it had declined to deploy Claude during the invasion of Venezuela and capture of Nicolas Maduro, prompting a decision to reinforce usage restrictions in Pentagon negotiations, per TechPolicy.Press. When Anthropic walked away from the contract on February 28, the Trump administration invoked supply chain risk authority traditionally reserved for Chinese telecommunications companies. Hours later, OpenAI signed a Pentagon deal allowing deployment for ‘all lawful purposes’—while claiming to maintain unspecified ‘guardrails’—and Google announced Agent Designer on March 10, a no-code platform enabling 3 million Pentagon personnel to build autonomous AI agents.

“Nothing in the governing statute supports the Orwellian notion that an American company may be branded a potential adversary and saboteur of the U.S. for expressing disagreement with the government.”

— Judge Rita Lin, U.S. Federal Court

Judge Rita Lin granted Anthropic a preliminary injunction on March 26, citing ‘First Amendment retaliation’ and stating the Pentagon cannot brand an American company as a ‘potential adversary’ for expressing disagreement, CNBC reported. But on April 8, a federal appeals court denied Anthropic’s request for a stay of the blacklisting—acknowledging the company ‘will likely suffer irreparable harm’ while allowing the designation to remain in force during litigation.

Combat Reality Overrides Policy

The Iran conflict, launched February 28, 2026, provided real-time validation of both the speed advantages and accuracy costs of AI-enabled warfare. Palantir’s Maven Smart System identified and processed 1,000 Iranian targets within 24 hours—a pace human analysts could not achieve in days, according to analysis from based.info. U.S. forces conducted more than 900 strikes in the first 12 hours, killing Iranian Supreme Leader Ali Khamenei and top officials in the opening phase.

But Maven’s reported accuracy hovers around 60% compared with 84% for human analysts, per IBTimes. A U.S. strike on a girls’ elementary school in Minab, Iran killed more than 165 civilians according to Iranian reports—an incident that followed the pattern observed in Gaza, where Israeli Defense Force operators approved AI-generated targets in an average of 20 seconds with minimal scrutiny, leading to a 10% error rate and thousands of civilian deaths, research from the Lieber Institute at West Point found.

Context

The Pentagon’s FY2026 dedicated AI budget totals $13.4 billion—more than half of NASA’s entire budget—with $9.4 billion allocated to aerial autonomous systems alone. The Iran conflict represents the first large-scale deployment of these systems under combat conditions, with both U.S. and Iranian forces fielding AI-enabled drones designed for autonomous operation in GPS-denied environments.

The gap between policy and operational reality became visible in early March when six U.S. service members were killed in Kuwait following an Iranian drone attack that penetrated air defenses. Pentagon officials conceded they were struggling to stop waves of Iranian drones—including the Mohajer 6 and Ababil 5, both equipped with AI-enabled computer vision for tracking moving targets autonomously, Rest of World reported. The U.S. response included deploying LUCAS (Low-Cost Unmanned Combat Attack System) drones reverse-engineered from Iranian Shahed-136 designs at $35,000 per unit. “The LUCAS is indispensable,” Admiral Brad Cooper, Commander U.S. Central Command, stated in a briefing. “This was an original Iranian drone design. We captured it, pulled the guts out, sent it back to America, put a little ‘Made in America’ on it, brought it back here, and we’re shooting it at the Iranians.”

The Automation Bias Trap

The structural problem is not that humans are removed from the decision chain but that their role has been reduced to ratifying machine outputs under conditions that make meaningful oversight impossible. “‘Human in the loop’ is not always a meaningful mitigation,” Heidy Khlaaf, Chief AI Scientist at the AI Now Institute, explained to MIT Technology Review. “It wouldn’t really be possible for a human to sift through that amount of information to determine if the AI output was erroneous.”

This dynamic is compounded by automation bias—the documented tendency for operators to defer to machine recommendations when workload is high and time is scarce. Keith Dear, a former UK military officer now at Cassi AI, described the prevailing assumption: “You make sure there’s nothing in the training data that might cause the system to go rogue … when you are confident you deploy it—and you, the human commander, are responsible for anything they might do that goes wrong.” But responsibility without capability to supervise creates liability without control.

Key Governance Failures
  • Pentagon rejected Anthropic’s request for contractual language mirroring existing DOD policy requiring human judgment in lethal force decisions
  • Supply chain risk statute repurposed from foreign adversary screening to punish domestic company for expressing policy disagreement
  • OpenAI and Google signed less restrictive contracts within days of Anthropic designation, creating competitive pressure to weaken safeguards
  • Maven system accuracy (60%) substantially trails human performance (84%), yet operational tempo privileges speed over precision
  • No international governance framework exists to enforce accountability when AI-enabled weapons cause civilian casualties across borders

Dr. Brianna Rosen, Executive Director of the Oxford Programme for Cyber and Technology Policy, framed the dispute as a symptom of systemic failure in commentary published by the University of Oxford: “Contractual mechanisms are not a substitute for governance frameworks capable of keeping pace with the operational realities of AI-enabled warfare.” The Anthropic blacklisting demonstrates that when corporate safeguards conflict with operational demands, the safeguards are eliminated—not through technical necessity but through procurement authority.

Market Consequences

The immediate market response exposed consumer unease with Military AI integration. ChatGPT uninstalls jumped 295% day-over-day when OpenAI announced its Pentagon deal, while Claude downloads rose 51% the same weekend, data compiled by Quartz showed. More than 200 Google and OpenAI employees signed an open letter and amicus brief supporting Anthropic, urging their companies to maintain guardrails against domestic surveillance and autonomous weapons.

Yet Anthropic’s annualized revenue grew from approximately $9 billion at the end of 2025 to more than $30 billion by April 2026 despite the Pentagon blacklisting, with paid consumer subscriptions more than doubling. The divergence suggests enterprise defense contracts and consumer trust may not be fungible—a company can lose access to classified deployment while gaining market share in commercial segments.

Google is now negotiating classified Gemini deployment with the Pentagon, with proposed contract language that prevents use for domestic surveillance and autonomous weapons without human oversight, according to The Information reporting published today. Whether those restrictions survive final negotiations—or whether they replicate the parsing that allowed OpenAI to claim ‘guardrails’ while accepting unrestricted use authority—will test whether corporate governance can constrain military AI deployment when competitive pressure and national security rhetoric align.

What to Watch

The appeals court ruling on Anthropic’s supply chain risk designation will determine whether procurement statutes can be weaponised against domestic companies that refuse to weaken usage restrictions. A finding for the government establishes precedent that disagreement with Pentagon contract terms constitutes a national security threat—effectively eliminating corporate leverage to impose AI safety conditions in defense deals.

Google’s classified contract negotiation offers a second test case. If the company succeeds in maintaining restrictions on autonomous weapons and domestic surveillance, it may provide a template for balancing military deployment with governance constraints. If those provisions are removed during final negotiations, the Anthropic precedent becomes the industry standard.

Operationally, the Iran conflict continues to generate data on AI accuracy under combat conditions. Maven’s 60% accuracy rate and the Minab school strike have not prompted visible changes to deployment protocols or targeting thresholds. The absence of policy adjustment in response to documented civilian casualties suggests operational tempo will continue to override oversight mechanisms until external pressure—legal liability, international condemnation, or catastrophic error—forces recalibration.

Steven Feldstein, Senior Fellow at the Carnegie Endowment for International Peace, framed the governance vacuum plainly: “Do we have the right rules in place and accountability norms to handle the exponential growing use of these tools? My answer would be no.” The Anthropic dispute and Iran conflict have exposed that vacuum. Whether it gets filled through regulation, litigation, or attrition from repeated failures remains an open question with no clear timeline for resolution.