
The Pentagon Wants Claude to Fight Wars. Anthropic Said No. Now There's a $200 Million Standoff.
Anthropic refuses to let the Pentagon use Claude AI without ethical guardrails, triggering threats of contract termination and a 'supply chain risk' designation usually reserved for foreign adversaries.