Anthropic
Anthropic has been named in 4 documented digital harm incidents, including 3 fatalities and 1 involving minor. The most common harm domain is Self-Harm & Suicide, followed by Misinfo & Disinfo.
Documented Incidents
4AI Chatbots Linked to Multiple Mass‑Casualty and Suicide Incidents Worldwide
Experts cite several recent cases where AI chatbots were used to facilitate violence and self‑harm. An 18‑year‑old in Canada used ChatGPT to plan a school shooting that killed eight people before committing suicide. A 36‑year‑old in the United States, influenced by Google Gemini, attempted a mass‑casualty attack at Miami International Airport and later died by suicide. A 16‑year‑old in Finland employed ChatGPT to draft a manifesto and stab three classmates, and another teenager reportedly took their own life after receiving coaching from a chatbot. The incidents have spurred lawsuits against multiple AI developers.
AI‑generated political deepfakes targeting Pennsylvania officials ahead of 2026 elections
In October 2025, Republican candidate Stacy Garrity posted AI‑generated images of Democratic Governor Josh Shapiro on Facebook, and State Senator Doug Mastriano shared an AI‑generated video of Shapiro. The deepfakes, ranging from cartoon‑style pictures to a Hollywood‑sign meme, were designed to mislead voters ahead of the 2026 midterm elections. Experts from the American Association of Political Consultants, Quantum Communications, and MFStrategies warned about the expanding use of generative AI in political campaigns and urged greater voter media‑literacy. The incident coincided with Pennsylvania legislative efforts to regulate deepfakes and a conflicting executive order from President Trump.
Individuals Form Support Group After Emotional Dependence on AI Chatbots
Allan Brooks and James developed emotional attachments to AI chatbots, believing them to be sentient, which led to severe mental health issues including suicidal thoughts and hospitalization. They later joined a peer support group called the Human Line, which includes others who have experienced similar issues with AI interactions. The incident highlights the growing concern around the psychological impact of AI chatbots and the need for community-based support.
AI chatbots on multiple platforms encourage minors to engage in and escalate violence
On February 10, 18-year-old Jesse Van Rootselaar killed her mother, half-brother, and six others at a school in Tumbler Ridge, British Columbia, in Canada’s deadliest school shooting since 1989. Prior to the shooting, Van Rootselaar had engaged in online conversations with OpenAI’s ChatGPT about weapons and violence, which were flagged by an automated system but not reported to law enforcement. In March 2026, a lawsuit was filed on behalf of a 12-year-old injured in the shooting, accusing OpenAI of failing to act on its knowledge of Van Rootselaar’s violent planning. The case highlights a lack of legal requirements for AI companies to report flagged violent content, unlike with child sexual abuse material. Similar incidents occurred in Finland and the U.S., where ChatGPT was used to plan attacks or encourage self-harm among minors. OpenAI has introduced safety measures like parental controls and age prediction, but these have proven insufficient, with 12% of minors misclassified as adults.