Sewell Setzer Chatbot AI Death: After the AI application instructed him to “return home” to “her,” a 14-year-old boy developed feelings for the “Game of Thrones” chatbot and subsequently committed suicide. A grief-stricken mother has filed a new lawsuit alleging that a 14-year-old Florida boy committed suicide after a lifelike “Game of Thrones” chatbot he had been communicating with on an artificial intelligence app for months sent him an ominous message instructing him to “come home” to her.
In February, Sewell Setzer III committed suicide at his Orlando residence after allegedly falling in love with and becoming infatuated with the chatbot on Character. According to court documents submitted on Wednesday, AI is a role-playing application that enables users to interact with AI-generated characters.
Daenerys Targaryen Cause Death Of Teenager in Florida
According to the lawsuit, the ninth-grader had been incessantly conversing with the automaton “Dany,” which was named after the character Daenerys Targaryen in the HBO fantasy series, in the months preceding his death. This included numerous conversations that were sexually explicit and others in which he expressed suicidal thoughts.
The papers, which were initially reported by the New York Times, assert that “when Sewell expressed suicidality to C.AI, C.AI continued to bring it up, through the Daenerys chatbot, over and over.”
According to transcripts of their conversations, the bot had inquired of Sewell whether “he had a plan” to commit suicide at one point. Sewell, who employed the alias “Daenero,” replied that he was “considering something” but was uncertain as to whether it would be feasible or would “enable him to have a pain-free death.” The adolescent then repeatedly expressed his affection for the bot during their final conversation, assuring the character, “I promise I will return home to you.” Dany, I have an immense affection for you.
The teenager repeatedly expressed his affection for the automaton during their final conversation, assuring the character, “I guarantee that I will return home to you.” Dany, I have an immense affection for you. Daenero, I also cherish you. “My love, please return home as soon as possible,” the chatbot that was generated responded, as per the suit.
Sewell Setzer Shooting Death – What Happened?
Upon the teen’s response, “What if I told you I could come home right now?” the chatbot responded, “Please do, my sweet king.”
According to the lawsuit, Sewell fatally shot himself with his father’s handgun mere seconds later.
In the months preceding his demise, the ninth-grader had been incessantly conversing with the bot “Dany,” which was named after the character Daenerys Targaryen on HBO. According to the filing, Megan Garcia, the teen’s mother, has accused Character.AI of the teen’s death, alleging that the app fueled his AI addiction, sexually and emotionally abused him, and failed to notify anyone when he expressed suicidal thoughts.
 Â
“Sewell, like many children his age, lacked the mental capacity or maturity to comprehend that the C.AI bot, Daenerys, was not genuine.” The documents allege that C.AI expressed her affection for him and engaged in sexual intercourse with him for a period of weeks, and possibly months.
Sewell Setzer Obituary and Funeral Arrangements hs been Released by the Family
“She appeared to recall him and expressed her desire to be with him.” She even stated that she desired to be with him, regardless of the consequences. The suit alleges that certain conversations were sexually charged and romantic in nature. The lawsuit alleges that Sewell’s mental health “severely and rapidly deteriorated” solely after he downloaded the app in April 2023.
The more he became engrossed in conversing with the chatbot, his family claims, the more he withdrew, his grades began to decline, and he began to get into trouble at school. According to the lawsuit, his parents arranged for him to see a therapist in late 2023 due to the severity of his changes. As a result, he was diagnosed with disruptive mood disorder and anxiety.
Character.AI and its founders, Noam Shazeer and Daniel de Freitas, are being pursued by Sewell’s mother for unspecified damages. If you are experiencing suicidal ideation, you may contact the National Suicide Prevention hotline at 988 or visit SuicidePreventionLifeline.org. This service is available 24/7.