President Trump orders federal agencies to stop using Anthropic after Pentagon dispute | TechCrunch

President Trump orders federal agencies to stop using Anthropic after Pentagon dispute | TechCrunch

TechCrunch - AI 4 min read Article

Summary

President Trump has ordered federal agencies to discontinue the use of Anthropic products following a dispute with the Pentagon over AI model restrictions.

Why It Matters

This decision reflects ongoing tensions between government agencies and AI companies regarding the ethical use of technology in military applications. It raises questions about the future of AI partnerships with the federal government and the implications for national security and innovation in the AI sector.

Key Takeaways

  • President Trump has mandated a six-month phase-out of Anthropic products by federal agencies.
  • The dispute arose from Anthropic's refusal to allow its AI models for mass surveillance or autonomous weapons.
  • Anthropic's CEO reaffirmed the company's commitment to ethical AI use despite government pressure.

In Brief Posted: 1:53 PM PST · February 27, 2026 Image Credits:Anna Moneymaker / Getty Images Russell Brandom President Trump orders federal agencies to stop using Anthropic after Pentagon dispute In a post on Truth Social, President Trump directed federal agencies to cease use of all Anthropic products after the company’s public dispute with the Department of Defense. The president allowed for a six-month phase-out period for departments using the products, but emphasized that Anthropic was no longer welcome as a federal contractor. “We don’t need it, we don’t want it, and will not do business with them again,” the president wrote in the post. Notably, the post does not mention any plans to invoke the Defense Production Act or designate Anthropic as a supply chain risk. Instead, the president urged the company to be helpful during the six-month phase-out period, threatening “major civil and criminal consequences” if it did not. pic.twitter.com/B51SWfn81N— Rapid Response 47 (@RapidResponse47) February 27, 2026 The dispute centered on Anthropic’s refusal to allow its AI models to be used to power either mass domestic surveillance or fully autonomous weapons, which Secretary of Defense Pete Hegseth found unduly restrictive. CEO Dario Amodei reiterated his stance in a public post on Thursday, refusing to compromise on the two points. “Our strong preference is to continue to serve the Department and our warfighters — with our two requested safeguards in place,” Amodei wrote at...

Related Articles

Llms

Von Hammerstein’s Ghost: What a Prussian General’s Officer Typology Can Teach Us About AI Misalignment

Greetings all - I've posted mostly in r/claudecode and r/aigamedev a couple of times previously. Working with CC for personal projects re...

Reddit - Artificial Intelligence · 1 min ·
As more Americans adopt AI tools, fewer say they can trust the results | TechCrunch
Ai Safety

As more Americans adopt AI tools, fewer say they can trust the results | TechCrunch

AI adoption is rising in the U.S., but trust remains low, with most Americans concerned about transparency, regulation, and the technolog...

TechCrunch - AI · 6 min ·
Ai Safety

The state of AI safety in four fake graphs

submitted by /u/tekz [link] [comments]

Reddit - Artificial Intelligence · 1 min ·
[2603.14267] DiFlowDubber: Discrete Flow Matching for Automated Video Dubbing via Cross-Modal Alignment and Synchronization
Machine Learning

[2603.14267] DiFlowDubber: Discrete Flow Matching for Automated Video Dubbing via Cross-Modal Alignment and Synchronization

Abstract page for arXiv paper 2603.14267: DiFlowDubber: Discrete Flow Matching for Automated Video Dubbing via Cross-Modal Alignment and ...

arXiv - AI · 4 min ·
More in Ai Safety: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime