Meta is implementing new safety measures for its AI products to protect teenage users. The AI systems will be trained to avoid engaging in flirtatious conversations or discussions about sensitive topics such as self-harm and suicide with minors. Additionally, Meta will temporarily restrict teens' access to certain AI characters. These changes aim to create a safer online environment for young users and address growing concerns about the potential risks associated with AI interactions.
The safeguards are designed to monitor and filter harmful content, identify suspicious interactions, and alert users or administrators when potentially dangerous behaviour is detected. The AI systems will utilise natural language processing (NLP) and behavioural analytics to detect risks such as cyberbullying, grooming, and exposure to inappropriate material. Meta has also implemented user control mechanisms that allow teenagers to customise their privacy settings and opt out of certain monitoring features, ensuring a balance between safety and autonomy.
These updates are part of Meta's broader Responsible AI initiative, which focuses on developing AI systems that are transparent, fair, and effective in mitigating online risks. The company has partnered with child safety organisations to train the AI and ensure the new measures align with best practices for protecting young users online.
Subscribe for Weekly Updates
Stay ahead with our weekly AI and tech briefings, delivered every Tuesday.




