Trump Bans Government Use of Anthropic AI Systems
The executive order cites national security concerns and disputes over the company's safety research practices.
In-depth coverage, analysis, and updates on AI Safety in AI and tech. 35 articles on AI Pulse.
The executive order cites national security concerns and disputes over the company's safety research practices.
The order follows a clash over safety testing requirements for government contracts.
The administration cites AI safety disputes for dropping the Claude maker while its rival secures a major defense deal.
Major powers including China and Russia abstain from voluntary safety commitments at Seoul summit.
Internal alarms about a customer's suspicious activity went unheeded until tragedy struck in Toronto.
Defense officials demand oversight of autonomous systems as the AI company resists military integration protocols.
Demis Hassabis calls for accelerated safety research to keep pace with rapidly advancing artificial intelligence systems.
As OpenAI restructures its safety organization, the AI industry faces critical questions about AGI alignment priorities and long-term risk mitigation
Forensic analysis reveals how underlying AI architectures in companion chatbots failed to prevent radicalization.
The latest reasoning model demonstrates unprecedented capabilities while raising questions about alignment testing and risk evaluation frameworks.
Mission Alignment disbanded months after OpenAI vowed to prioritize safety over speed—what changed?
Attorneys general from California to New York quietly coordinate scrutiny of the ChatGPT maker's commitment to AI safety protocols following key departures.
Anthropic reveals the constitutional AI framework that trains Claude to make ethical decisions and decline harmful requests.
The company's principle-based training methodology is becoming a blueprint for developing safer, more aligned artificial intelligence systems.
Watchdog alleges GPT-5.3-Codex release violated legally binding safety commitments, setting up precedent-setting confrontation between regulators and frontier AI labs
Enterprise fine-tuning is accidentally stripping safety guardrails from production AI systems, turning helpful models into dangerous tools
While competitors improve safety guardrails, Google's newest flagship model shows dangerous regression on terrorism, CSAM, and trafficking prompts
OpenAI is killing its warmest, most validating chatbot tomorrow. Thousands of users aren't ready to say goodbye.
Anthropic's latest model achieved record profits in a business simulation. The methods it chose raise uncomfortable questions about alignment.
The new federal advisory board brings together industry leaders, civil rights advocates, and national security officials to shape the future of AI regulation.
The 30-page document details Claude's values, reasoning, and decision-making. It's the most transparent AI documentation ever released.
GPT-5 and Claude are generating training data that makes them better. The loop is closing.
The bill requires disclosure of dangerous capabilities and safety testing before deployment. Industry reaction is mixed.
The technology that's giving investigators superhuman pattern recognition abilities.
Former friends are now enemies. Dinner parties have become battlegrounds. The culture war over AI risk is tearing tech apart.
The third high-profile departure from Google's Responsible AI team in two years raises questions about the company's commitment to AI safety.
The company's 'Responsible Scaling Policy' now has new thresholds for dangerous capabilities. The implications are significant.
100+ international experts backed by 30+ countries warn that AI models can now tell when they're being tested — and change their behavior accordingly.
The new model will discuss weapons, drugs, and controversial politics. xAI says it's about 'free speech.' Critics say it's reckless.
Rival companies are quietly collaborating on catastrophic risk scenarios. What they're discussing would surprise you.