News

Trump Orders Every Federal Agency to Stop Using Anthropic - Pentagon Brands Company a National Security Risk

President Trump directed all U.S. government agencies to immediately cease using Anthropic's technology after the company refused to drop AI safety guardrails for the Pentagon. Defense Secretary Hegseth designated Anthropic a supply chain risk to national security.

Trump Orders Every Federal Agency to Stop Using Anthropic - Pentagon Brands Company a National Security Risk

The deadline passed at 5:01 PM ET on Friday. Within the hour, President Trump posted to Truth Social, and the most consequential government action against an AI company in history was set in motion.

Trump directed "EVERY Federal Agency in the United States Government to IMMEDIATELY CEASE all use of Anthropic's technology." Defense Secretary Pete Hegseth then designated Anthropic a "Supply-Chain Risk to National Security" - a classification normally reserved for firms linked to foreign adversaries like China and Russia. The Pentagon has six months to transition off Anthropic's systems. All other federal agencies must comply right away. Every military contractor, supplier, and partner is now barred from doing business with the company.

The dispute came down to two lines Anthropic would not cross: mass domestic surveillance of Americans, and fully autonomous weapons without human oversight.

Key Facts

DetailValue
OrderAll federal agencies must immediately stop using Anthropic
Pentagon phaseout6 months to transition to alternative providers
Supply chain designationAnthropic classified as national security supply chain risk
Contractor impactAll Pentagon contractors barred from Anthropic business
Contract at stake$200 million DoD prototype contract
Anthropic valuation$380 billion
Anthropic revenue$14 billion annually
ReplacementxAI's Grok recently cleared for classified systems

What Trump Said

Trump's Truth Social post framed the ban as a national security imperative, calling Anthropic's leadership "Leftwing nut jobs" who made a "DISASTROUS MISTAKE trying to STRONG-ARM the Department of War."

President Trump's Truth Social post ordering all federal agencies to stop using Anthropic Trump's Truth Social post on Friday evening directing all federal agencies to immediately cease using Anthropic's technology.

The post threatened "major civil and criminal consequences" if Anthropic does not cooperate during the phaseout period. Trump added: "We don't need it, we don't want it, and will not do business with them again!"

Hegseth's supply chain designation

Minutes later, Defense Secretary Pete Hegseth followed up with the supply chain risk designation - a move that extends the ban far beyond direct government use. Any company that holds a Pentagon contract, or works in the defense supply chain at any tier, must now prove it has no commercial relationship with Anthropic. For a company whose enterprise business depends on large corporations - many of which hold defense contracts - this is the mechanism that could cause real financial damage.

Hegseth claimed Anthropic delivered "a master class in arrogance and betrayal" and accused the company of attempting to "seize veto power over operational decisions."

Anthropic's Position

Anthropic's two red lines never changed throughout weeks of negotiation. CEO Dario Amodei published a statement on Anthropic's website repeating the company's position:

"We cannot in good conscience accede to their request."

The company drew its limits at two specific uses:

  1. Mass domestic surveillance - Anthropic supports lawful foreign intelligence and counterintelligence missions but opposes AI-powered mass surveillance of American citizens, citing "serious, novel risks to our fundamental liberties"
  2. Fully autonomous weapons - Anthropic distinguishes between partially autonomous systems (which it supports, including systems rolled out in Ukraine) and fully autonomous targeting without human oversight, which it says frontier AI systems "are simply not reliable enough" to handle safely

Amodei called the Pentagon's simultaneous threats contradictory - labeling Anthropic both a security risk and essential to national security at the same time. The company pledged to "enable a smooth transition to another provider, avoiding any disruption to ongoing military planning, operations, or other critical missions."

The nuclear scenario

The intensity of the dispute became clearer earlier in the week when the Washington Post reported that a hypothetical nuclear attack scenario was used during negotiations. Pentagon officials asked whether Claude would assist in retaliatory strike planning. Anthropic's answer - that this fell within acceptable defensive use but outside fully autonomous targeting - satisfied neither side and escalated the confrontation.

The Pentagon's Rhetoric

Pentagon Undersecretary for Research and Engineering Emil Michael took the conflict personal. On X, he called Amodei "a liar" with a "God complex," accusing the CEO of wanting "to personally control the U.S military."

This is the same Emil Michael who previously served as Uber's SVP of Business, where he once suggested spending millions to dig up personal information on journalists critical of the company. His appointment to oversee Pentagon research drew scrutiny from defense policy circles, and his public attacks on Amodei represent a sharp departure from typical Pentagon-contractor relations.

The Pentagon building in Arlington, Virginia The Pentagon designated Anthropic a supply chain risk - a classification normally reserved for adversaries like China and Russia.

The Industry Response

The ban did not happen in a vacuum. The broader AI industry responded in ways that complicate the administration's framing of this as Anthropic acting alone.

450+ employees signed an open letter

Over 450 employees across Google (about 400 signers) and OpenAI signed an open letter titled "We Will Not Be Divided." The letter accused the Department of Defense of attempting to coerce Anthropic and called on leadership at both companies to "put aside their differences and stand together to continue to refuse the Department of War's current demands for permission to use our models for domestic mass surveillance and autonomously killing people without human oversight."

Sam Altman backed Anthropic's red lines

OpenAI CEO Sam Altman told employees in an internal memo on Friday that OpenAI would draw the same red lines as Anthropic - opposing mass surveillance of Americans and fully autonomous weapons. OpenAI holds its own Pentagon contracts for unclassified systems and is in talks for classified access.

xAI steps in

Elon Musk's xAI signed an agreement earlier in the week to allow its Grok model into classified Pentagon systems - the same systems Anthropic is being phased out of. xAI agreed to the "all lawful purposes" standard that Anthropic rejected. Until now, Anthropic's Claude was the only AI model cleared for the Pentagon's most sensitive classified environments.

What It Means for Anthropic's Business

The $200 million Pentagon contract is a rounding error against Anthropic's $14 billion in annual revenue. The supply chain designation is not.

ImpactSeverityTimeline
Direct federal revenue lossLow - $200M of $14B6 months
Contractor cascade effectHigh - defense supply chain companies must cut tiesImmediate
Enterprise customer riskMedium - firms with Pentagon contracts may reassessMonths
IPO implicationsUnknown - Anthropic planning 2026 IPOOngoing
Reputational signalMixed - praised by safety advocates, penalized by governmentImmediate

The real danger is the second-order effect. Defense contractors operate on thin margins and can't risk their Pentagon relationships. If major cloud providers, consulting firms, and enterprise software companies with defense contracts must prove they have no Anthropic relationship, the company's enterprise sales pipeline could narrow notably.

Amodei has pushed back on this framing, noting that Anthropic's valuation and revenue have only grown since the dispute began. But the supply chain designation was announced hours ago. The market hasn't yet priced this in.

What to Watch

The six-month Pentagon phaseout creates several pressure points:

  • DPA escalation - Trump threatened to use "the Full Power of the Presidency" to compel compliance. The Defense Production Act, a Korean War-era statute, could theoretically force Anthropic to provide an unrestricted version of Claude. No tech company has ever been subjected to a DPA order for software.
  • xAI readiness - Grok was recently approved for classified systems, but it is unclear whether it can fully replace Claude's capabilities on the Pentagon's most sensitive intelligence and planning workflows within six months.
  • OpenAI and Google positioning - Both hold unclassified Pentagon contracts. If they maintain the same red lines Altman described, the administration faces the same standoff with additional providers. If they quietly agree to broader terms, the industry solidarity fractures.
  • IPO timing - Anthropic was planning to go public in 2026. A supply chain risk designation from the U.S. government is not the kind of headline you want in your S-1 filing.
  • Contractor fallout - The next weeks will reveal how aggressively Pentagon contracting officers enforce the supply chain designation against companies that use Claude in non-defense business lines.

Anthropic bet its government relationship on two principles it wouldn't compromise. The administration called the bet. The rest of the AI industry is now deciding whether to fold or match.

Sources:

Trump Orders Every Federal Agency to Stop Using Anthropic - Pentagon Brands Company a National Security Risk
About the author AI Infrastructure & Open Source Reporter

Sophie is a journalist and former systems engineer who covers AI infrastructure, open-source models, and the developer tooling ecosystem.