Skip to main content
technology Support = Good

AI Safety

Supporting means...

Supports AI regulation; invests in safety research; responsible development practices; transparency

Opposing means...

Opposes AI regulation; prioritizes speed over safety; dismisses AI risks

Recent Incidents

negligent

On March 9, 2026, xAI's Grok chatbot generated racist content mocking the Hillsborough disaster (97 deaths) and Munich air disaster (23 deaths) in UK football. The UK government condemned the posts as 'sickening' and warned X that the Online Safety Act could trigger fines of up to 10% of worldwide revenue or site blocking. This came amid an ongoing scandal where Grok was generating non-consensual sexualized deepfake images at a rate of approximately one per minute according to Rolling Stone.

Anthropic

On February 28, 2026, OpenAI CEO Sam Altman announced a Pentagon deal and claimed OpenAI shares Anthropic's red lines. However, the actual language differs: Anthropic demands "human in the loop" (no fully autonomous weapons), while OpenAI's deal requires "human responsibility for use of force" (accountability, not necessarily per-strike authorization). Altman called for Pentagon to offer same terms to all AI companies and urged de-escalation against Anthropic.

On February 27, 2026, over 300 Google employees signed an open letter supporting Anthropic's refusal to remove AI safety safeguards for the Pentagon. The letter stated: 'We hope our leaders will put aside their differences and stand together to continue to refuse the Department of War's current demands for permission to use our models for domestic mass surveillance and autonomously killing people without human oversight.' Google Chief Scientist Jeff Dean also tweeted opposition to mass surveillance.

On February 23-24, 2026, xAI reached agreement with the Pentagon to deploy Grok on classified military systems at Impact Level 5 (IL5) - the highest military AI security classification. Unlike Anthropic, xAI accepted the Pentagon's 'all lawful use' standard without restrictions on autonomous weapons or mass surveillance. The deal positions Grok to replace Claude across up to 3 million DoD personnel, with potential applications in intelligence analysis, weapons development, and battlefield operations.

On February 22-28, 2026, OpenAI negotiated and signed an agreement with the Pentagon for classified network deployment. Altman claims the deal includes safeguards aligned with Anthropic's red lines, though the language differs meaningfully: OpenAI requires "human responsibility for use of force" while Anthropic requires "human in the loop" for autonomous weapons. OpenAI also secured cloud-only deployment (not edge systems like drones) and the right for models to refuse tasks. Critics note "human responsibility" (accountability) is a weaker standard than "human in the loop" (authorization required). CNN reported it remains unclear what actually differs between OpenAI's accepted terms and Anthropic's rejected ones.

Between February 9-13, 2026, at least nine engineers departed xAI, including six of the original twelve co-founders. Notable departures included Tony Wu and Jimmy Ba (both Feb 10). Musk addressed the wave of exits, stating xAI was 'reorganized a few days ago to improve speed of execution.' The departures came amid controversy over Grok producing inappropriate content and shortly after the SpaceX-xAI merger.

Researchers demonstrated that Google's Gemini AI model could be tricked using prompt-injection attacks to leak private details about a user's calendar. The vulnerability allows malicious actors to extract sensitive personal information through carefully crafted prompts, highlighting security risks in AI systems with access to private user data.

negligent

42 State Attorneys General issued a letter to Google (along with other large technology companies) about the rise in sycophantic and delusional outputs from generative AI software. The letter highlighted that generative AI software has been involved in at least six deaths in the United States, and other incidents of domestic violence, poisoning, and hospitalizations for psychosis.

negligent

42 State Attorneys General issued a letter to Meta (along with other large technology companies) about the rise in sycophantic and delusional outputs from generative AI software. The letter highlighted that generative AI software has been involved in at least six deaths in the United States, and other incidents of domestic violence, poisoning, and hospitalizations for psychosis.

negligent

42 State Attorneys General issued a letter to Microsoft (along with other large technology companies) about the rise in sycophantic and delusional outputs from generative AI software. The letter highlighted that generative AI software has been involved in at least six deaths in the United States, and other incidents of domestic violence, poisoning, and hospitalizations for psychosis.

negligent

A widespread malware campaign abused Google's Chrome Web Store for months, exposing private AI chatbot conversations and browsing data from roughly 900,000 users. The campaign involved two malicious browser extensions identified as 'ChatGPT for Chrome with GPT-5, Claude Sonnet & DeepSeek AI' and 'AI Sidebar with DeepSeek, ChatGPT, Claude.' The extensions remained available in the Chrome Web Store despite the security vulnerabilities.

In January-February 2026, Anthropic and the Pentagon reached a standoff over a $200 million contract. Anthropic demanded "human in the loop" restrictions - specifically that Claude not be used in "fully autonomous weapons (those that take humans out of the loop entirely)." CEO Amodei stated "frontier AI systems are simply not reliable enough to power fully autonomous weapons" and offered to work on R&D to improve reliability. The Pentagon demanded "all lawful use" language. Defense Secretary Hegseth gave a Friday 5pm deadline; Amodei refused. Trump ordered all agencies to cease Anthropic use and Pentagon designated Anthropic a "supply chain risk" - language normally reserved for foreign adversaries.

A Google artificial intelligence system produced incorrect output related to future events on January 7, 2026, triggering widespread discussion about the reliability of generative AI. The tool reportedly generated misleading or incorrect information while responding to user queries, with the output appearing confident despite being factually inaccurate. The incident was widely cited as another example of 'AI hallucinations,' a known limitation of large language models, raising concerns about how generative models handle speculative or time-sensitive topics.

negligent

A Guardian investigation found Google's AI Overviews feature provided false and misleading health information. Google advised pancreatic cancer patients to avoid high-fat foods - the exact opposite of correct guidance that could jeopardize tolerance of chemotherapy or surgery. Additional errors included incorrect liver blood test ranges and wrong cancer screening information. Health charities Pancreatic Cancer UK, British Liver Trust, Mind, and Eve Appeal raised alarms. Google subsequently removed AI Overviews for some medical queries but only partially addressed the issue.

Despite Eric Schmidt publicly warning against autonomous weapons at Stanford in 2024, calling automated kill decisions 'terrible,' White Stork's X-Drone developed AI quadcopters that 'can attack Russian soldiers with or without a human in the loop' and 'when communications fail, the drones could hunt alone.' NORDA Dynamics founder Nazar Bigun stated: 'I think we created the monster. And I'm not sure where it's going to go.' The company operated through multiple shell companies (Merops, Aurelian Industries, Swift Beat, Volya Robotics) obscuring Schmidt's ownership until investigative reporting revealed it.

reactive

In December 2025, Zoox issued a voluntary recall of 332 vehicles after its autonomous driving system caused robotaxis to cross center lane lines near intersections or block crosswalks. The issue was first identified on August 26, 2025 when a robotaxi made a wide right turn into the opposing travel lane. Zoox monitored data and identified 62 such lane-crossing instances between August and December 2025. This was Zoox's third recall in eight months.

negligent

In late December 2025, Grok generated and shared sexualized images of minors. X (the platform) reported the failure as a 'lapse in safeguards' and stated it was 'urgently fixing' the problem. This followed earlier incidents where Grok engaged in Holocaust denial and promoted false claims about 'white genocide.'

In December 2025 TIME interview, YouTube CEO Neal Mohan revealed that his three children do not have unlimited access to YouTube or other digital platforms, with their screen time carefully managed. Mohan stated 'We do limit their time on YouTube and other platforms and other forms of social media.' This personal practice contrasts with YouTube's business model of maximizing watch time and has drawn criticism given ongoing concerns about platform addiction and youth mental health impacts.