Anthropic sues Defense Department over supply chain risk designation


Anthropic has made good on its promise to challenge the Department of Defense in court after the agency labeled it a supply chain risk late last week.

The Claude-maker filed a complaint against the Department on Monday. The complaint comes after a weeks-long conflict between Anthropic and the DOD over whether the military should have unrestricted access to Anthropic’s AI systems. Anthropic had two firm red lines: it didn’t want its technology to be used for mass surveillance of Americans and didn’t believe it was ready to power fully autonomous weapons with no humans making targeting and firing decisions.

Defense Secretary Pete Hegseth argued that the Pentagon should have access to AI systems for “any lawful purpose.” A supply chain risk label is usually reserved for foreign adversaries, and requires any company or agency that does work with the Pentagon to certify that it doesn’t use Anthropic’s models. 

Anthropic called the DOD’s actions “unprecedented and unlawful” in a complaint filed in San Francisco federal court. “The Constitution does not allow the government to wield its enormous power to punish a company for its protected speech.”

This story is developing. Please check back for updates.

Anthropic to challenge DOD’s supply-chain label in court


Dario Amodei said Thursday that Anthropic plans to challenge the Department of Defense’s decision to label the AI firm a supply-chain risk in court, a designation he has called “legally unsound.”

The statement comes a few hours after the DOD officially designated Anthropic a supply-chain risk following a weeks-long dispute over how much control the military should have over AI systems. A supply-chain risk designation can bar a company from working with the Pentagon and its contractors. Amodei drew a firm line that Anthropic’s AI will not be used for mass surveillance of Americans or for fully autonomous weapons, but the Pentagon believed it should have unrestricted access for “all lawful purposes.”

In his statement, Amodei said the vast majority of Anthropic’s customers are unaffected by the supply-chain risk designation.

“With respect to our customers, it plainly applies only to the use of Claude by customers as a direct part of contracts with the Department of War, not all use of Claude by customers who have such contracts,” he said.

As a preview of what Anthropic will likely argue in court, Amodei said the Department’s letter labeling the firm a supply-chain risk is narrow in scope.

“It exists to protect the government rather than to punish a supplier; in fact, the law requires the Secretary of War to use the least restrictive means necessary to accomplish the goal of protecting the supply chain,” Amodei said. “Even for Department of War contractors, the supply chain risk designation doesn’t (and can’t) limit uses of Claude or business relationships with Anthropic if those are unrelated to their specific Department of War contracts.”

Amodei reiterated that Anthropic had been having productive conversations with the DOD over the last several days, conversations that some suspect got derailed when an internal memo he sent to staff was leaked. In it, Amodei characterized rival OpenAI’s dealings with the Department of Defense as “safety theater.”

Techcrunch event

San Francisco, CA
|
October 13-15, 2026

OpenAI has signed a deal to work with the DOD in Anthropic’s place, a move that has sparked backlash among OpenAI staff.

Amodei apologized for the leak in his Thursday statement, claiming that the company did not intentionally share the memo or direct anyone else to do so. “It is not in our interest to escalate the situation,” he said.

Amodei said the memo was written within “a few hours” of a series of announcements, including a presidential Truth Social post saying Anthropic would be removed from federal systems, then Defense Secretary Pete Hegseth’s supply-chain risk designation, and finally the Pentagon’s deal announcement with OpenAI. He apologized for the tone, calling it “a difficult day for the company” and said the memo didn’t reflect his “careful or considered views.” Written six days ago, he added, it’s now an “out-of-date assessment.”

He finished by saying Anthropic’s top priority is to ensure American soldiers and national security experts maintain access to important tools in the middle of ongoing major combat operations. Anthropic is currently supporting some of the U.S.’s operations in Iran, and Amodei said the company would continue to provide its models to the DOD at “nominal cost” for “as long as necessary to make that transition.”

Anthropic could challenge the designation in federal court, likely in Washington, but the law behind the decision makes it harder to contest because it limits the usual ways companies can challenge government procurement decisions and gives the Pentagon broad discretion on national security matters.

Or as Dean Ball — a former Trump-era White House adviser on AI who has spoken out against Hegseth’s treatment of Anthropic — put it: “Courts are pretty reluctant to second-guess the government on what is and is not a national security issue … There’s a very high bar that one needs to clear in order to do that. But it’s not impossible.”