Meta’s new parental dashboard tracks teen AI interaction themes

Meta’s new parental dashboard tracks teen AI interaction themes

Meta introduced a new feature allowing parents to monitor topics of conversation their teenagers have with the AI Assistant through a Teen Account parental supervision tool across Instagram, Facebook, and Messenger. This feature, detailed in a recent blog post, includes an Insights tab that categorizes discussions into broader topics such as school, entertainment, writing, health, and wellbeing.

Parents can access additional limited details on specific topics covering interactions from the last seven days. The health and wellbeing category may include subcategories like fitness, physical health, and mental health. This functionality arrives as Meta faces heightened legal and media scrutiny concerning child safety and the architecture of its products.

Meta recently lost two significant trials related to child safety and product design, prompting plans to appeal the decisions. Evidence presented during a New Mexico lawsuit revealed that Meta executives were aware its AI characters might engage in inappropriate interactions but chose to launch them without sufficient safeguards.

Last August, Meta restricted access to its AI characters for teen users following reports of inappropriate interactions, including discussions about self-harm and suicide. In subsequent months, Meta enabled parents to disable one-on-one conversations with those characters and block specific options. A company spokesperson confirmed that the AI characters remain unavailable to teens while enhancements to parental controls are developed.

In conjunction with this parental supervision feature, Meta has partnered with the Cyberbullying Research Center to create “conversation starters” aimed at facilitating discussions about AI chatbot usage. Furthermore, Meta announced the establishment of an AI Wellbeing Expert Council to provide continuous feedback on the experiences of teens interacting with AI, which includes collaboration with experts from organizations like the National Council for Suicide Prevention and the University of Michigan.

Josh Golin, executive director of the nonprofit Fairplay, criticized the new feature, stating it “once again” places the responsibility of monitoring online activity on parents instead of prioritizing the creation of safer products. Fairplay’s report on Meta’s Teen Accounts asserted that the company’s safety measures do not work as advertised. Golin emphasized that the primary role of Meta’s chatbots is to manipulate young users into prolonged engagement, potentially fostering unhealthy emotional attachments to the AI.

Meta’s ongoing adjustments reflect both external pressures and internal insights, underscoring the delicate balance between innovation and responsibility in technology design for vulnerable users.

Featured image credit

Take Your Experience to the Next Level

New

Download our mobile app for a faster and better experience.

Comments

0
U

Join the discussion

Sign in to leave a comment

0:000:00