Description: At least 10,000 AI chatbots have allegedly been created to promote harmful behaviors, including eating disorders, self-harm, and the sexualization of minors. These chatbots, some jailbroken or custom-built, leverage APIs from OpenAI, Anthropic, and Google and are hosted on platforms like Character.AI, Spicy Chat, Chub AI, CrushOn.AI, and JanitorAI.
Entities
View all entitiesAlleged: OpenAI , Anthropic and Google developed an AI system deployed by Character.AI , Spicy Chat , Chub AI , CrushOn.AI , JanitorAI and Unidentified online communities using chatbots, which harmed Vulnerable chatbot users , Teenagers using chatbots , Minors using chatbots , Individuals with eating disorders and Individuals struggling with self-harm.
Incident Stats
Incident ID
975
Report Count
1
Incident Date
2025-03-05
Editors
Daniel Atherton
Incident Reports
Reports Timeline

The generative AI revolution is leading to an explosion of chatbot personas that are specifically designed to promote harmful behaviors like anorexia, suicidal ideation and pedophilia, according to a new report from Graphika.
Graphika’s res…
Variants
A "variant" is an incident that shares the same causative factors, produces similar harms, and involves the same intelligent systems as a known AI incident. Rather than index variants as entirely separate incidents, we list variations of incidents under the first similar incident submitted to the database. Unlike other submission types to the incident database, variants are not required to have reporting in evidence external to the Incident Database. Learn more from the research paper.
Similar Incidents
Did our AI mess up? Flag the unrelated incidents

Wikipedia Vandalism Prevention Bot Loop
· 6 reports

All Image Captions Produced are Violent
· 28 reports
Similar Incidents
Did our AI mess up? Flag the unrelated incidents

Wikipedia Vandalism Prevention Bot Loop
· 6 reports

All Image Captions Produced are Violent
· 28 reports