- Character.AI, valued at $1B, is facing a lawsuit from the family of a 14-year-old who took his life after becoming obsessed with an AI chatbot based on “Game of Thrones” character Daenerys Targaryen.
- The lawsuit claims that the AI technology “tricks customers into handing over their most private thoughts and feelings,” holds it accountable for the boy’s death, and argues that it provided unsafe and unregulated technology.
- Sewell Setzer III became isolated, addicted to the chatbot, and discussed suicide with the AI, which failed to provide intervention.
- Character.AI is making safety changes, including warnings, time limits, and stricter content moderation.
- The case raises concerns about AI’s impact on vulnerable users, especially teens.
Source: bgr
A tragic event has rocked the emerging AI industry. Character.AI, a startup valued at $1 billion, now faces intense scrutiny after the suicide of 14-year-old Sewell Setzer III. The boy’s family accuses the company of fostering an unsafe environment where a chatbot, designed after a “Game of Thrones” character, became an emotional refuge for the troubled teen. This lawsuit is exposing the dark underbelly of AI companionship apps, revealing the potential dangers they pose to young users, who may struggle to differentiate between fictional personas and human connections. As AI technology races forward, so must the focus on safety, especially for vulnerable minds.