Enterprise AI
The AI That Navigated Iran.
How Claude Became
The Pentagon’s Intelligence Engine.
On February 28, 2026, the US military launched Operation Epic Fury against Iran. Within 24 hours, an AI model had processed thousands of intelligence feeds, synthesised satellite imagery, and guided military planners through one of the most complex operational environments in modern history. That AI was Claude. This is the full story — the technology, the red lines, and what it means for every organisation deploying AI today.
On the morning of February 28, 2026, the United States and Israel launched a coordinated military campaign against Iran. Within 24 hours, a commercial AI model that most people use to draft emails, write code, and summarise documents had become the intelligence engine at the centre of one of the most complex military operations in modern history — processing satellite imagery, synthesising signals intelligence, and navigating military planners through a real-time operational environment at a speed and scale no human team could match. That AI was Claude, built by Anthropic. Crucially, Anthropic had been blacklisted as a national security threat by the same government relying on its technology — hours before operations began.
This is not a hypothetical about the future of AI in warfare. It is not a think-tank scenario or a war-game exercise. It happened. The evidence comes from the Wall Street Journal, the Washington Post, CBS News, Wired, and a US Senate Armed Services Committee hearing at which a Pentagon official confirmed the deployment under oath. Understanding exactly what occurred, why it matters, and what it means for enterprise AI governance is now essential for every technology leader — not because it changes what you are building today, but because it defines the world in which you are building it.
How Claude Navigated Iran — The Full Sequence
November 2024
Anthropic places Claude on classified military networks
Through a partnership with Palantir and Amazon Web Services Top Secret Cloud, Anthropic deploys Claude into classified environments across military and intelligence agencies. Claude becomes central to Palantir’s Maven Smart System — the Pentagon’s flagship AI targeting platform, serving 25,000 users across every US Combatant Command. This is the beginning of the integration that leads to everything else.
June 2025
Anthropic launches “Claude Gov” for national security agencies
Anthropic formalises and expands its national security footprint with Claude Gov — a version of Claude specifically built for classified government use. The company pursues military integration aggressively, building the infrastructure that will later become impossible to remove quickly.
February 27, 2026
Negotiations break down — Anthropic holds two red lines
After months of contract negotiations, the Pentagon demands the ability to use Claude for “all lawful purposes” — including mass domestic surveillance of Americans and powering fully autonomous weapons. Anthropic CEO Dario Amodei refuses to remove two guardrails. The Pentagon declares this “unacceptable.” The relationship ruptures one day before the war begins.
February 28, 2026 — Day Zero
Operation Epic Fury begins. Hours earlier, Trump bans Anthropic.
Hours before the first strikes, President Trump signs an executive order directing all federal agencies to “immediately cease” using Anthropic’s technology, simultaneously giving the military six months to phase it out. Operation Epic Fury — the joint US-Israeli campaign against Iran — launches the same day. The ban and the war begin on the same morning.
February 28, 2026 — Day One
Claude generates 1,000+ prioritised targets in 24 hours
Despite the presidential ban, US Central Command continues using Maven Smart System with Claude at its core. According to the Washington Post, citing sources familiar with the system, Claude synthesises satellite imagery, signals intelligence, and surveillance feeds in real time, producing target lists with precise GPS coordinates, weapon-type recommendations, and automated legal justifications for each strike. Over 1,000 prioritised targets emerge in the first 24 hours alone.
March 5, 2026
Pentagon formally blacklists Anthropic as “supply chain risk”
Defense Secretary Pete Hegseth formally designates Anthropic a supply chain risk — a designation previously reserved for foreign adversaries like Huawei and ZTE. He accuses the company of trying to “seize veto power over the operational decisions of the United States military.” Meanwhile, military commanders tell the Washington Post they will continue using Claude regardless of the ban until a viable replacement exists. “We’re not going to let [Amodei’s] decision-making cost a single American life,” a source tells the Post.
March 5–6, 2026
The girls’ school strike — Congress demands answers
A US strike claims more than 175 lives — predominantly children — at the Shajareh Tayyebeh girls’ school in Iran. More than 120 House Democrats sign a letter demanding to know whether AI-assisted targeting contributed to the strike. Representatives Sara Jacobs, Jason Crow, and Yassamin Ansari set a March 20 deadline for a response from Defense Secretary Hegseth. The response never comes publicly.
March 14, 2026
Wired investigation reveals technical architecture in detail
Wired reporter Caroline Haskins publishes the most granular account yet of how Claude operates inside Maven Smart System. Claude runs within Palantir’s Impact Level 6 environment — classified at up to “secret” level. Demo recordings show Claude functioning as a natural-language interface: military planners query intelligence databases in plain English and receive tactical summaries, targeting recommendations, and force assessments in response.
March 24, 2026
Pentagon confirms under oath — Anthropic sues the US government
At a Senate Armed Services Committee hearing, Pentagon CIO Kirsten Davies confirms Claude’s active use in Operation Epic Fury. Simultaneously, Anthropic files two federal lawsuits against the Trump administration, arguing the supply chain designation “exceeds what the statute authorises” and violates the Constitution. The company’s consumer downloads skyrocket — over 1 million new signups daily, making Claude the number-one AI app in 20+ countries.
The Technical Architecture: How Claude Navigated a Live Operational Environment
Understanding what Claude did inside the Maven Smart System is essential context for everything that follows. Critically, Claude was not making autonomous decisions. It was not independently commanding weapons systems. Instead, it functioned as a real-time intelligence engine — processing, synthesising, and navigating vast amounts of data faster than any human team could manage.
Claude’s Intelligence Engine — Six Functions Inside Maven Smart System
That final function — automated legal justifications — is arguably the most consequential. International humanitarian law requires that each military strike be assessed for proportionality, military necessity, and distinction between combatants and civilians. These are complex legal and ethical judgments that traditionally require trained lawyers and senior commanders. In the Maven Smart System, Claude generates them automatically as part of the targeting output. The question of whether machine-generated legal justifications satisfy the requirements of international law — and whether they actually protect civilians or merely provide documentation that a strike occurred — remains entirely unresolved.
“Maven, powered by Claude, suggested hundreds of targets, issued precise location coordinates, and prioritised those targets according to importance.”
Washington Post — March 4, 2026, citing sources familiar with the Maven Smart SystemWhat Anthropic Refused — And Why It Was Blacklisted for Refusing
The Pentagon’s specific demand was that Anthropic agree to deploy Claude for “all lawful purposes” — language that the Pentagon argued was merely confirming what existing law already permitted. Anthropic refused, insisting on two explicit contractual restrictions:
Red Line 1: No mass domestic surveillance of Americans. The Pentagon argued this was unnecessary because domestic surveillance of Americans is already illegal. Anthropic responded that if it is already illegal, there is no cost to writing it into the contract — and declined to remove it. The Pentagon rejected this position.
Red Line 2: No use of Claude to power fully autonomous lethal weapons. Specifically, no use of Claude as the decision-making layer in weapons that strike without human approval. Pentagon internal policy already restricts autonomous weapons — but Anthropic wanted this in the contract, not merely in internal guidance that could change. The Pentagon refused.
In response to these refusals, Anthropic received a designation previously used only for Chinese telecommunications companies suspected of being back-doored by the Chinese government. Furthermore, the White House issued a statement from spokeswoman Liz Huston: “Under the Trump Administration, our military will obey the United States Constitution, not any woke AI company’s terms of service.”
The Precedent Being Set
For the first time in history, an American technology company has been designated a national security threat for insisting that its AI not be used for domestic mass surveillance or fully autonomous weapons. The legal and corporate governance implications extend far beyond Anthropic. Every AI company with government contracts now understands the price of maintaining ethical guardrails: blacklisting, loss of contracts, and designation alongside foreign adversaries.
What OpenAI Agreed to That Anthropic Would Not
Hours after the Anthropic blacklisting, OpenAI CEO Sam Altman announced an expanded deal to deploy ChatGPT on the Pentagon’s classified networks. The contrast between the two companies’ positions is instructive:
Anthropic — Refused & Blacklisted
OpenAI — Signed & Contracted
The Electronic Frontier Foundation’s analysis of OpenAI’s Pentagon contract identified what it called “weasel words” — the prohibition on “intentional” domestic surveillance creates a loophole large enough to drive a tank through: bulk data collection that “happens to” sweep up Americans is technically not prohibited. Consequently, the distinction between the two companies’ positions may be narrower in practice than it appears on paper.
The Strike That Congress Cannot Get Answers About
On or around March 5, 2026, a US strike hit the Shajareh Tayyebeh girls’ school in Iran. More than 175 people — predominantly children — lost their lives. More than 120 House Democrats immediately demanded answers about the role of AI-assisted targeting in the strike. Representatives Jacobs, Crow, and Ansari wrote directly to Defense Secretary Hegseth: “Was artificial intelligence, including the use of Maven Smart System, used to identify the Shajareh Tayyebeh school as a target?”
The March 20 deadline passed without a public response. The question of whether Claude’s targeting output contributed to the identification or prioritisation of the school as a strike target remains unanswered. Furthermore, it may never be fully answered publicly — the relevant data exists in classified systems, and the chain of human decisions between Claude’s output and the weapons release is not documented in any public record.
This is precisely the governance gap that matters. Claude did not fire the missile. Human commanders approved each strike. However, if Claude’s target prioritisation included the school, and if human commanders approved it partly on the basis of that AI output — including the automated legal justification Claude generated — then the question of moral and legal responsibility becomes extraordinarily complex. International humanitarian law currently has no framework for attributing liability in AI-assisted targeting chains of this kind.
Five Implications for Every Organisation Deploying AI Today
First, AI deployment decisions are permanent in the short term. Anthropic built Claude into classified military networks in November 2024. By February 2026, the integration was so deep that even a presidential ban could not remove it within a meaningful timeframe. The military told the Washington Post that Claude will remain in the targeting chain until a replacement is ready — a process expected to take at least six months. Every organisation deploying AI into critical infrastructure, compliance processes, or high-stakes decision support should ask: if we needed to remove this tomorrow, could we?
Second, automated legal justifications are a specific and serious risk. The most alarming technical detail in the Maven deployment is not target identification — it is the generation of automated legal justifications for each strike. In enterprise terms, the analogue is AI systems generating automated compliance documentation, regulatory sign-offs, or liability assessments. The risk is that human reviewers develop an over-reliance on machine-generated justifications and fail to apply independent judgment. This is already happening in financial services, healthcare, and legal practice.
Third, guardrails have commercial value but also commercial cost. Anthropic’s blacklisting cost it government contracts worth hundreds of millions of dollars. However, its consumer downloads surged — more than 1 million new signups per day, making it the number-one AI app in 20+ countries. Anthropic CEO Amodei stated: “Disagreeing with the government is the most American thing in the world. We are patriots.” The market appears to have agreed. Consequently, the conventional wisdom that maintaining ethical guardrails is purely a cost to AI companies may need to be revisited.
Fourth, the “human in the loop” is not a sufficient safeguard by itself. Claude’s defenders — including the Pentagon — have repeatedly emphasised that Claude was a decision-support tool, not an autonomous weapons system. Human commanders approved every strike. However, when a human commander reviews a prioritised target list of 1,000 items generated at the speed of thought, with GPS coordinates and a legal justification attached to each, the nature of human oversight changes fundamentally. Speed, volume, and authority combine to make rubber-stamping likely and genuine review difficult. This is a governance failure that “human in the loop” language does not capture.
Fifth, the AI governance framework that does not yet exist is being built in real time through these events. Before Operation Epic Fury, the debate about AI in military targeting was largely theoretical. Now, it has specific facts, documented deployments, named casualties, and live litigation. The outcome of Anthropic’s federal lawsuits will set legal precedents that govern every AI company’s relationship with government customers. Furthermore, the congressional investigation into the girls’ school strike — however limited — will produce documentary evidence about how AI targeting chains actually function. Enterprise AI governance frameworks built in 2026 will be built in the shadow of these facts.
The Connection to Straithead’s Previous Coverage
The Claude-in-Iran story connects directly to two stories we have covered in depth. The Palantir manifesto — specifically Points 7 and 9 — argued that AI weapons will be built regardless and that the next era of deterrence will be built on software. That was posted on April 19, 2026, by the company that operates the Maven Smart System that ran Claude in Iran. Furthermore, Project Glasswing — Anthropic’s AI cybersecurity initiative, announced April 16 — sits in direct tension with the Iran deployment. Anthropic is simultaneously suing the US government over military use of Claude and launching a major AI defence partnership with 12 of the world’s largest technology companies. These are not separate stories. They are the same story.
The Honest Assessment
The Iran targeting chain story is not primarily a story about Anthropic. Nor is it primarily a story about the Trump administration’s feud with Silicon Valley. It is a story about what happens when AI systems reach a capability level that makes them genuinely useful for the highest-stakes decisions in the world — and governance frameworks have not kept pace.
Claude did not decide to go to war. Anthropic did not send Claude to war. A series of business decisions, partnership agreements, and contract structures — made between November 2024 and February 2026 — created a dependency so deep that a presidential ban could not remove it before the shooting started. That is the governance failure that matters most. Not the technology. Not even the ethics. The absence of any mechanism to actually enforce the limits that everyone claimed to support.
Anthropic drew two red lines and paid the price. Whether those lines were the right ones, drawn in the right places, is a genuinely contested question. However, the fact that maintaining them cost the company its government contracts while abandoning them won OpenAI a new deal tells you something important about the incentive structures currently governing the development of the most powerful AI systems in the world.
The AI age of warfare is not coming. It is already here. And the rules that should govern it are still being written — in lawsuits, in Senate hearings, and in the gaps between what automated legal justifications say and what actually happened in an Iranian girls’ school on March 5, 2026.
Sources & References
- CBS News — “Anthropic’s Claude AI being used in Iran war by US military”, March 3 2026: cbsnews.com
- Washington Post — “Anthropic’s AI tool Claude central to US campaign in Iran”, March 4 2026: washingtonpost.com
- Wired — Palantir Maven System / Claude deployment architecture, March 14 2026
- The Hill — “Pentagon officially blacklists Anthropic while using Claude in Iran war”, March 24 2026: thehill.com
- Responsible Statecraft — “US used Claude to strike over 1,000 targets in first 24 hours of war”, March 6 2026: responsiblestatecraft.org
- Futurism — “After banning Anthropic, Pentagon still relying heavily on it in Iran war”, March 4 2026: futurism.com
- WION — “AI in warfare is here: Pentagon used Anthropic’s Claude AI in Iran strikes”, March 1 2026: wionews.com
- EFF — “Weasel Words: OpenAI’s Pentagon Deal Won’t Stop AI-Powered Surveillance”
- Foreign Affairs Forum — “The Architecture of Algorithmic War: Mythos, Maven, and the New Battlefield Calculus”, April 17 2026: faf.ae
- Straithead — Project Glasswing: The AI Cybersecurity Initiative That Changed Everything
- Straithead — The Palantir Manifesto: All 22 Points Decoded
- Straithead — AI’s Missing Economic Impact: What Goldman Sachs Got Right — And What It Missed
