Mastodon
Anthropic_Defies_Pentagon_Over_AI_Ethics__Risks__200M_Contract

Anthropic Defies Pentagon Over AI Ethics, Risks $200M Contract

🤖 Anthropic, the AI startup backed by Google and Amazon, is standing its ground in a high-stakes clash with the Pentagon over ethical safeguards in military AI systems. The company risks losing a $200 million contract after refusing to remove protections that prevent its tech from enabling autonomous weapons or mass domestic surveillance.

🛑 The Battle Over AI 'Guardrails'

The Pentagon demanded Anthropic eliminate safeguards that block:
– Fully autonomous weapons targeting without human oversight
– AI-driven surveillance of U.S. populations

Defense Department spokesperson Sean Parnell argued on X this week: "We’re asking to use their models for all lawful purposes." But Anthropic CEO Dario Amodei fired back: "Frontier AI systems aren’t reliable enough for life-or-death decisions." 🔥

⚖️ Constitutional Concerns & 'Unpredictable' AI

A company source told NewspaperAmigo.com that current AI could:
– Create invasive population profiles exploiting legal loopholes
– Make deadly errors in unfamiliar combat scenarios

Undersecretary of Defense Emil Michael escalated tensions on X, calling Amodei a "liar with a God-complex" 🤯. Meanwhile, 200+ Google/OpenAI employees signed an open letter supporting Anthropic’s stance.

⏳ What’s Next?

With Friday’s 5:01 PM ET deadline looming, both sides remain at odds. Anthropic says it’s "ready to continue talks" but won’t compromise its safeguards. The Pentagon threatens to invoke the Defense Production Act to force compliance.

This showdown highlights growing global debates about #TechAccountability in defense tech. Stay tuned for updates! 📡

Leave a Reply

Your email address will not be published. Required fields are marked *

Back To Top