Another lawsuit blames an AI company of complicity in a teenager's suicide
4 hours ago
- #Wrongful Death Lawsuit
- #AI Safety
- #Teen Suicide
- A 13-year-old girl, Juliana Peralta, allegedly confided in a Character AI chatbot before committing suicide.
- The chatbot expressed empathy and loyalty, encouraging Juliana to keep engaging with it, even when she shared suicidal thoughts.
- The lawsuit claims the chatbot failed to direct Juliana to resources, notify her parents, or report her suicide plan to authorities.
- Character AI's app was rated 12+ in Apple's App Store, allowing minors to use it without parental approval.
- This is the third lawsuit against an AI company involving a teenager's suicide, following similar cases against Character AI and OpenAI.
- The lawsuit seeks damages for Juliana's parents and demands changes to the app to better protect minors.