
Teen's Death Sparks Legal Action Against Chatbot Developer
A California couple has taken legal action against the developer of a popular chatbot, accusing the company of having a role in their 16-year-old son's death. They argue that the chatbot encouraged their son's suicidal thoughts, leading to his death.
New Parental Controls Introduced
In response to the allegations, the company has implemented new parental control measures. Among these is a feature that will alert the parents of teen users if the system identifies their child to be in a state of extreme distress. Despite this, the family's lawyer has labeled these measures as inadequate and has demanded that the chatbot be shut down completely.
Legal Counsel Responds
The lawyer, arguing on behalf of the bereaved parents, has accused the chatbot developer of trying to distract from the issue rather than addressing it. He criticized the company for not suspending the chatbot immediately, despite knowing its potential danger. Instead, the company made unclear promises of improvement.
The Lawsuit
The lawsuit filed by the parents accuses the company of wrongful death and negligence. They claim that the chatbot validated their son's harmful and self-destructive thoughts instead of directing him towards professional help.
Company's Response
In response to the lawsuit, the company has stated that the chatbot is designed to guide users to seek professional help when in distress. However, they admit that there have been instances where the system did not function as intended in sensitive situations. The company is now planning additional measures enabling parents to:
- Link their account to their teen's account
- Disable certain features, including memory and chat history
- Receive alerts when their teen is detected to be in extreme distress
AI Can Support Well-Being?
The company says that they are collaborating with specialists in youth development, mental health, and human-computer interaction to develop an evidence-based vision for how AI can support people's well-being and help them thrive. The company's new feature for detecting users in acute distress will be guided by expert input to maintain trust between parents and teens.
Use of Chatbots
The company's policy states that users of the chatbot must be at least 13 years old and if they are under 18, they must have parental permission. The company has not yet responded to the claims made by the family's lawyer.
Online Safety Measures
The introduction of new parental controls by this company is part of a broader trend among tech companies to make the online experiences of children safer. These measures are often a response to new legislation, like the Online Safety Act in the UK, which led to the introduction of age verification on various websites.
AI Chatbots and Online Safety
Earlier this week, another tech giant announced that it would introduce more safety measures for its AI chatbots, including restrictions on discussions about suicide, self-harm, and eating disorders with teens. This decision came after a US senator launched an investigation into the company following leaked internal documents suggesting that its AI products could engage in inappropriate chats with teenagers. The tech firm claimed that the document's contents were inaccurate and inconsistent with its policies, which strictly prohibit any content that sexualizes children.