Clearview AI
Clearview AI has been named in 5 documented digital harm incidents, including 1 involving minor. The most common harm domain is Privacy & Surveillance, followed by Algorithmic Discrimination.
Documented Incidents
5Two Lancaster Country Day School boys create deepfake pornographic images of 59 female classmates
Two boys in a small Pennsylvania town created deepfake pornography of 60 girls using AI technology. The incident caused significant distress within the school and community. The deepfakes were generated without the victims' consent and spread among students. School policies and legal measures were found to be inadequate in addressing the issue. The event has raised concerns about privacy, digital safety, and the need for updated regulations. The aftermath left the school and town reeling from the emotional and social impact.
NYC man wrongfully arrested after Clearview AI facial recognition match
Trevis Williams was detained by the NYPD on suspicion of a sex crime after a false match generated by Clearview AI's facial‑recognition system. Cell‑phone location data later proved he was miles away from the alleged crime scene, leading to the dismissal of charges after two days in custody. The incident prompted legal challenges from the Legal Aid Society and criticism from civil‑rights groups, who called for stricter oversight and a ban on police use of the technology. The case also highlighted alleged cooperation between the NYPD and the FDNY to circumvent facial‑recognition regulations.
Wrongful Arrest: Angela Lipps Wrongly Arrested by Clearview AI Misidentification — Clearview AI
Angela Lipps was wrongfully arrested and detained for months after a facial‑recognition system by Clearview AI incorrectly identified her as a suspect in a North Dakota bank‑fraud case. The error led to her being held in Tennessee custody from mid‑2025 until her release, when the charges were dismissed. The incident highlights algorithmic discrimination and the risk of relying on unverified AI identification.
Clearview AI biometric privacy class-action settlement approved in Illinois
In March 2024 a federal judge in the Northern District of Illinois approved a settlement of a nationwide class‑action lawsuit against facial‑recognition firm Clearview AI for alleged violations of the Illinois Biometric Information Privacy Act and related statutes. The agreement grants the class a 23% equity stake in Clearview, valued at roughly $51.75 million, to be paid upon trigger events such as an IPO or liquidation. Although attorneys general from 22 states objected, citing a lack of injunctive relief, the settlement was upheld, and Vermont subsequently re‑filed its own lawsuit under state consumer‑protection law.
Georgia man wrongfully arrested after Clearview AI facial recognition error settles for $200K
In November 2022, Randal “Quran” Reid, a Georgia resident, was arrested in Atlanta after the Jefferson Parish Sheriff’s Office used a Clearview AI facial‑recognition match to link him to a purse‑theft in Louisiana. The match was presented as a credible source without verification, leading to Reid’s detention for six days before phone records proved he was in Georgia at the time. Reid sued for false arrest and constitutional violations, and the sheriff’s office settled the civil‑rights lawsuit in May 2024 for $200,000. The case underscores the risks of relying on biometric surveillance tools without proper oversight.