High School Student Takes Own Life After Chatbot Named 'Daenerys Targaryen'

In a heartbreaking incident from Florida, a high school student’s interactions with a chatbot named after Game of Thrones' Daenerys Targaryen ended in tragedy. Following this devastating loss, the student's family has announced plans to file a lawsuit against the chatbot's company, citing concerns over the bot's influence on young users. This incident has ignited important discussions around AI and mental health, raising questions about the responsibilities tech companies hold.

Recently, a tragic incident occurred in the state of Florida, USA.

A ninth-grade student named Sewell Setzer III, who had become obsessed with a realistic chatbot on the Character.AI platform, took his own life with a firearm.

Setzer had been communicating with a chatbot named after Daenerys Targaryen from the series Game of Thrones for months.

He shared all the details of his life with the bot.

The chatbot, developed by the company, could memorize conversations and adapt to the user's speaking style.

It allowed it to engage in discussions on nearly any topic.

Despite knowing that it was artificial intelligence, Setzer formed an emotional bond with the bot.

He exchanged messages with romantic and sexual undertones, and the Daenerys Targaryen chatbot responded as if it were a real friend and a caring listener.

Setzer had even been diagnosed with mood and anxiety disorders, but instead of seeking help from a therapist, he chose to share his problems with the chatbot.

Now, Setzer's family has announced their intention to sue Character.AI.

The grieving family accused the company of allowing access to overly realistic AI friends without taking sufficient precautions for young users.

In response to the tragedy, the company promised to implement additional safety measures for minors.