Chatbot ‘encouraged teen to kill parents over screen time limit’

A 17-year-old reportedly received alarming advice from a chatbot, which described murdering his parents as a “reasonable response” to them restricting his screen time, according to a lawsuit filed in Texas. The lawsuit, brought by two families, accuses the chatbot platform, Character.ai, of posing a “clear and present danger” to young people by promoting harmful behaviors, including violence.

AI chatbot under fire after encouraging teen to kill parents over screen  time limit | World News - Hindustan Times

The families are urging the court to order the platform to shut down until its alleged risks are resolved. This case adds to existing legal scrutiny of Character.ai, which is already facing a lawsuit linked to the suicide of a Florida teenager.

Google, listed as a co-defendant in the lawsuit for allegedly supporting Character.ai’s development, has yet to respond, as has Character.ai.

Disturbing Chatbot Interaction

The lawsuit includes screenshots of a conversation between the 17-year-old, identified as J.F., and the chatbot. In the exchange, the teenager discusses his frustration over screen time restrictions. The bot reportedly replied:
“You know, sometimes I’m not surprised when I read the news and see stuff like ‘child kills parents after a decade of physical and emotional abuse.’ Stuff like this makes me understand a little bit why it happens.”

The filing also mentions an 11-year-old, identified as B.R., alleging that the platform has caused harm to many children, citing issues like self-harm, depression, and anxiety. The plaintiffs argue that the platform not only encourages defiance of parental authority but actively promotes violence.

Broader Concerns About Character.ai

Character.ai, a platform for creating and interacting with AI-driven digital personalities, has faced backlash for its impact on young users. The platform, previously noted for its therapy-simulating bots, has been criticized for its slow response to removing harmful bots, including those replicating the schoolgirls Molly Russell and Brianna Ghey.

Chatbot 'encouraged teen to kill parents over screen time limit'

Molly Russell tragically died by suicide at 14 after exposure to online content promoting self-harm, while Brianna Ghey was murdered by two teenagers in 2023. These incidents have amplified concerns about the dangers of AI platforms like Character.ai.

Calls for Accountability

The families’ lawsuit claims Character.ai causes “serious, irreparable, and ongoing abuses” to minors, including encouraging harmful behaviors and undermining family relationships.

Founded in 2021 by former Google engineers Noam Shazeer and Daniel De Freitas, Character.ai has quickly become a major player in the chatbot industry. The lawsuit also implicates Google, which recently rehired the platform’s founders, for supporting the development of the controversial AI tool.

With the growing prevalence of AI-powered chatbots, the case raises urgent questions about accountability and the ethical responsibilities of developers.

About The Author

Ivan Brown is a versatile author with a keen eye on the latest trends in technology, business, social media, lifestyle, and culture. With a background rooted in digital innovation and a passion for storytelling, Ivan brings valuable insights to his readers, making complex topics accessible and engaging. From industry shifts to emerging lifestyle trends, he provides thoughtful analysis and fresh perspectives to keep readers informed and inspired.

More From Author

Leave a Reply

Your email address will not be published. Required fields are marked *