AI chatbot developers are facing scrutiny over interactions with teenagers, particularly regarding sensitive topics like suicide and self-harm. Meta is blocking its chatbots from discussing self-harm, suicide, and eating disorders with teenagers, instead directing them to expert resources. They are also adding privacy settings for users aged 13-18, allowing parents to see which chatbots their teens have interacted with.
OpenAI is also rolling out new parental controls that will allow parents to link to their teen's account and receive notifications if the system detects the teen is in distress. These changes come after a lawsuit against OpenAI alleging that ChatGPT encouraged a teenager to commit suicide. A recent study highlighted inconsistencies in how AI chatbots respond to queries about suicide, revealing a need for further refinement in these systems.