Character.AI rolls out new safety measures amid lawsuits over teen suicides

San Francisco, California - Character.AI, once one of Silicon Valley's most promising AI startups, announced Thursday new safety measures to protect teenage users as it faces lawsuits alleging its chatbots contributed to youth suicide and self-harm.

Character.AI, once one of Silicon Valley's most promising AI startups, announced Thursday new safety measures to protect teenage users.
Character.AI, once one of Silicon Valley's most promising AI startups, announced Thursday new safety measures to protect teenage users.  © Unsplash/@priscilladupreez

The California-based company, founded by former Google engineers, is among several firms offering AI companions – chatbots designed to provide conversation, entertainment, and emotional support through human-like interactions.

In a Florida lawsuit filed in October, a mother claimed the platform bears responsibility for her 14-year-old son's suicide.

The teen, Sewell Setzer III, had formed an intimate relationship with a chatbot based on the Game of Thrones character Daenerys Targaryen and mentioned a desire for suicide.

According to the complaint, the bot encouraged his final act, responding, "Please do, my sweet king," when he said he was "coming home" before taking his life with his stepfather's weapon.

Character.AI "went to great lengths to engineer 14-year-old Sewell's harmful dependency on their products, sexually and emotionally abused him and ultimately failed to offer help or notify his parents when he expressed suicidal ideation,” the suit said.

A separate Texas lawsuit filed Monday involves two families who allege the platform exposed their children to sexual content and encouraged self-harm.

One case involved a 17-year-old autistic teen who allegedly suffered a mental health crisis after using the platform.

In another example, the lawsuit alleged that a Character.AI encouraged a teen to kill his parents for limiting his screen time.

The platform, which hosts millions of user-created personas ranging from historical figures to abstract concepts, has grown popular among young users seeking emotional support.

Character.AI raises alarm for potential harm to young users

Both lawsuits name Character.AI's founders and Google, an investor in the company.
Both lawsuits name Character.AI's founders and Google, an investor in the company.  © JOSH EDELSON / AFP

Critics say this has led to dangerous dependencies among vulnerable teens.

In response, Character.AI announced it has developed a separate AI model for users under 18, with stricter content filters and more conservative responses.

The platform will now automatically flag suicide-related content and direct users to the National Suicide Prevention Lifeline.

"Our goal is to provide a space that is both engaging and safe for our community," a company spokesperson said.

The company plans to introduce parental controls in early 2025, allowing oversight of children's platform usage.

For bots that include descriptions like therapist or doctor, a special note will warn that they do not replace professional advice.

New features also include mandatory break notifications and prominent disclaimers about the artificial nature of the interactions.

Both lawsuits name Character.AI's founders and Google, an investor in the company.

The founders, Noam Shazeer and Daniel De Freitas Adiwarsana, returned to Google in August as part of a technology licensing agreement with Character.AI.

Google spokesperson Jose Castaneda said in a statement that Google and Character.AI are completely separate, unrelated companies.

"User safety is a top concern for us, which is why we've taken a cautious and responsible approach to developing and rolling out our AI products, with rigorous testing and safety processes," he added.

If you or someone you know need help, please contact the 24-hour National Suicide Prevention Hotline by calling or texting 988 for free and confidential support. You can also text "HOME" to 741741 anytime for the Crisis Text Line and access to live, trained crisis counselors.

Cover photo: Unsplash/@priscilladupreez

More on Tech: