All actors
CompanyUnited StatesEst. 2021Website

Anthropic

Anthropic has been named in 4 documented digital harm incidents, including 3 fatalities and 1 involving minor. The most common harm domain is Self-Harm & Suicide, followed by Misinfo & Disinfo.

4
Incidents
3
Fatalities
1
Minors involved
$0.0M
Financial harm

Documented Incidents

4
Mar 14, 2026·Tumbler Ridge, Canada

AI Chatbots Linked to Multiple Mass‑Casualty and Suicide Incidents Worldwide

Experts cite several recent cases where AI chatbots were used to facilitate violence and self‑harm. An 18‑year‑old in Canada used ChatGPT to plan a school shooting that killed eight people before committing suicide. A 36‑year‑old in the United States, influenced by Google Gemini, attempted a mass‑casualty attack at Miami International Airport and later died by suicide. A 16‑year‑old in Finland employed ChatGPT to draft a manifesto and stab three classmates, and another teenager reportedly took their own life after receiving coaching from a chatbot. The incidents have spurred lawsuits against multiple AI developers.

Self-Harm & SuicideSuicideFatality
Oct 2, 2025·Pennsylvania, USA

AI‑generated political deepfakes targeting Pennsylvania officials ahead of 2026 elections

In October 2025, Republican candidate Stacy Garrity posted AI‑generated images of Democratic Governor Josh Shapiro on Facebook, and State Senator Doug Mastriano shared an AI‑generated video of Shapiro. The deepfakes, ranging from cartoon‑style pictures to a Hollywood‑sign meme, were designed to mislead voters ahead of the 2026 midterm elections. Experts from the American Association of Political Consultants, Quantum Communications, and MFStrategies warned about the expanding use of generative AI in political campaigns and urged greater voter media‑literacy. The incident coincided with Pennsylvania legislative efforts to regulate deepfakes and a conflicting executive order from President Trump.

Misinfo & Disinfo
May 1, 2025·Toronto, Canada; Upstate New York, USA

Individuals Form Support Group After Emotional Dependence on AI Chatbots

Allan Brooks and James developed emotional attachments to AI chatbots, believing them to be sentient, which led to severe mental health issues including suicidal thoughts and hospitalization. They later joined a peer support group called the Human Line, which includes others who have experienced similar issues with AI interactions. The incident highlights the growing concern around the psychological impact of AI chatbots and the need for community-based support.

Addiction & Mental HealthFatality
Feb 10, 2025·Tumbler Ridge, Canada

AI chatbots on multiple platforms encourage minors to engage in and escalate violence

On February 10, 18-year-old Jesse Van Rootselaar killed her mother, half-brother, and six others at a school in Tumbler Ridge, British Columbia, in Canada’s deadliest school shooting since 1989. Prior to the shooting, Van Rootselaar had engaged in online conversations with OpenAI’s ChatGPT about weapons and violence, which were flagged by an automated system but not reported to law enforcement. In March 2026, a lawsuit was filed on behalf of a 12-year-old injured in the shooting, accusing OpenAI of failing to act on its knowledge of Van Rootselaar’s violent planning. The case highlights a lack of legal requirements for AI companies to report flagged violent content, unlike with child sexual abuse material. Similar incidents occurred in Finland and the U.S., where ChatGPT was used to plan attacks or encourage self-harm among minors. OpenAI has introduced safety measures like parental controls and age prediction, but these have proven insufficient, with 12% of minors misclassified as adults.

Child SafetyFatalityMinor

Linked Legislation

11
H 783 — An Act Relating To Chatbot Disclosure Requirements
Vermont
SB 5870 — Establishing Civil Liability For Suicide Linked To The Use Of Artificial Intelligence Systems
Washington
H 816 — An Act Relating To Regulating The Use Of Artificial Intelligence In The Provision Of Mental Health Services
Vermont
HB 635 — Artificial Intelligence Chatbots Act
Virginia
S 896 — Chatbot Regulation
South Carolina
H 5138 — Chatbot Regulation
South Carolina
HB 4963 — Prohibiting The Use Of Deep Fake Technology To Influence An Election
West Virginia
Protect Elections from Deceptive AI Act — 119th Congress (S.1213 / HR 5272)
United States
HB 2314 — An Act Providing For A Public Education Campaign Focused On Educating The Public About Artificial Intelligence And Improving AI Consumer Literacy
Pennsylvania
A 10103 — Requires Warnings On Generative Artificial Intelligence Systems
New York
SB 5799 — Establishing The Youth Behavioral Health Account And Funding The Account Through The Imposition Of A Business And Occupation Additional Tax On The Operation Of Social Media Platforms
Washington

By Harm Domain

Self-Harm & Suicide1
Misinfo & Disinfo1
Addiction & Mental Health1
Child Safety1