# AI safety
Latest news and articles about AI safety
Total: 8 articles found

Musk Opens Grok 4.2 Candidate to Public Beta, Promising Weekly ‘Fast‑Learning’ Updates
Elon Musk has opened a candidate public beta of Grok 4.2, requiring users to opt in and inviting public feedback. The model claims a new fast‑learning capability and will receive weekly updates accompanied by release notes, accelerating xAI’s iterative development approach but raising questions about safety and oversight.

OpenAI Recruits Creator of OpenClaw, Vows to Keep Viral Agent Open-Source via New Foundation
OpenAI has hired Peter Steinberger, creator of the widely adopted agent framework OpenClaw, and pledged to place the project into a foundation that will keep it open-source and independent while receiving funding and support. The move is a tactical win for OpenAI but raises questions about governance, security and the balance between openness and centralization as agent platforms mature.

US Military Allegedly Used Anthropic’s Claude in Venezuela Operation, Raising Questions About AI’s Role in War
U.S. media report that Anthropic’s AI model Claude was used in the January 3 U.S. operation in Venezuela, routed via a partnership with Palantir. Anthropic has not confirmed the claim and stresses its policy forbidding uses that facilitate violence, but the allegation raises legal, ethical and strategic questions about private AI models in military operations.

OpenAI’s Voice Models Tapped for Pentagon Drone‑Swarm Challenge, Raising Dual‑Use Concerns
OpenAI has shared an open‑source voice‑to‑instruction model with two Pentagon‑selected defence firms competing in a prize to produce voice‑controlled drone‑swarm prototypes. The move highlights the tension between commercial AI innovation and the risks of rapid diffusion of components that can enable more autonomous and potentially weaponised systems.

AI Insiders Sound the Alarm as U.S. Start‑ups Pivot from Safety to Speed
Senior researchers exiting US AI companies have publicly warned that commercialization and IPO pressures are sidelining safety, risking manipulative or harmful model behaviour. The conflict between monetisation incentives and the need for interpretability, privacy safeguards and robust alignment work has produced real‑world moderation failures and could invite regulatory intervention.

Musk’s AI Project in Retreat: Key xAI Founders Exit After SpaceX Rescue
Two prominent xAI founders quit within 48 hours after a series of earlier exits left half the original founding team gone, undermining Elon Musk’s AI ambitions. The exits, heavy cash burn, and product scandals around Grok have coincided with xAI’s absorption into SpaceX — a deal that looks like a financial bailout but raises fresh strategic and regulatory headaches.

OpenClaw’s Viral Rise Signals a New Age for Cheap, Deployable AI Agents — and New Risks
OpenClaw, an open‑source agent platform created by Peter Steinberger, has gone viral by turning chat messages into executable commands across multiple model APIs, accelerating demand for inexpensive, high‑throughput models and simple local hardware like the Mac Mini. The surge highlights opportunities for Chinese model providers such as Minimax and Kimi, while raising acute security, deployment and governance challenges.

Philippines to Lift Ban on xAI’s Grok After Promised Fixes for Sexual-Content Abuse
The Philippines will lift its ban on xAI’s Grok once the company implements promised fixes to stop the chatbot being used to generate sexually explicit images, including alleged child-exploitative content. Authorities will continue close monitoring, following platform-level restrictions introduced earlier by X to block generation of real-person nudity.