Meta is implementing new safety measures for its AI products to protect teenage users. The AI systems will be trained to avoid engaging in flirtatious conversations or discussions about sensitive topics such as self-harm and suicide with minors. Additionally, Meta will temporarily restrict teens' access to certain AI characters. These changes aim to create a safer online environment for young users and address growing concerns about the potential risks associated with AI interactions.
The safeguards are designed to monitor and filter harmful content, identify suspicious interactions, and alert users or administrators when potentially dangerous behaviour is detected. The AI systems will utilise natural language processing (NLP) and behavioural analytics to detect risks such as cyberbullying, grooming, and exposure to inappropriate material. Meta has also implemented user control mechanisms that allow teenagers to customise their privacy settings and opt out of certain monitoring features, ensuring a balance between safety and autonomy.
These updates are part of Meta's broader Responsible AI initiative, which focuses on developing AI systems that are transparent, fair, and effective in mitigating online risks. The company has partnered with child safety organisations to train the AI and ensure the new measures align with best practices for protecting young users online.
Related Articles
Microsoft salaries exposed amid talent war
Read more about Microsoft salaries exposed amid talent war →Meta Pauses AI Expansion
Read more about Meta Pauses AI Expansion →Brazil Orders Meta Action
Read more about Brazil Orders Meta Action →Meta AI: Restructuring Again
Read more about Meta AI: Restructuring Again →