Microsoft has taken an unprecedented public stand for Anthropic, filing a court brief in a San Francisco federal court that sends a clear warning to the Pentagon that AI ethics cannot be coerced through government designations and contract cancellations. The brief called for a temporary restraining order against the supply-chain risk designation and was joined by a separate filing from Amazon, Google, Apple, and OpenAI. The coordinated industry response is being described as the most significant legal pushback against a government action in the modern AI era.
The Pentagon’s designation was applied to Anthropic after the company refused to allow its Claude AI to be used for mass surveillance of US citizens or to power autonomous lethal weapons as part of a $200 million contract negotiation. Defense Secretary Pete Hegseth formalized the designation, triggering the cancellation of Anthropic’s government contracts. Anthropic filed two lawsuits in response, challenging the designation in both California and Washington DC courts on the same day.
Microsoft’s unprecedented stand is grounded in its direct integration of Anthropic’s technology into federal military systems and its participation in the Pentagon’s $9 billion cloud computing contract. The company also holds additional agreements with government agencies. Microsoft publicly called for a collaborative framework in which government and industry work together to ensure that AI serves national security without crossing ethical lines related to surveillance or unauthorized warfare.
Anthropic’s court filings argued that the supply-chain risk designation was an unconstitutional act of retaliation for the company’s publicly held AI safety positions. The company disclosed that it does not currently believe Claude is safe or reliable enough for lethal autonomous operations, which it said was the genuine technical basis for the restrictions it sought in the contract. The Pentagon’s technology chief publicly ruled out any possibility of renewed negotiations.
Congressional Democrats have separately asked the Pentagon whether AI was involved in a strike in Iran that reportedly killed more than 175 civilians at a school, raising concerns about AI targeting and human oversight. Their formal inquiries are adding legislative pressure to an already extraordinary legal confrontation. Together, Microsoft’s unprecedented stand, the industry coalition, and congressional scrutiny are forcing the Pentagon to publicly defend its approach to AI governance in warfare.
Microsoft Takes Unprecedented Stand for Anthropic, Warning Pentagon That AI Ethics Cannot Be Coerced
Date:
Picture Credit: Rawpixel (Public Domain)
