Meta Platforms is stepping up efforts to protect younger users on Instagram by introducing special teen accounts with enhanced privacy settings. This new feature, announced on Tuesday, marks Meta's latest move to reduce teens' exposure to harmful content across its platforms. The initiative comes in response to growing regulatory pressure, particularly around online safety.
Meta's update includes automatically converting all designated teen accounts to private accounts by default. With these changes, only followers or pre-connected users will be able to send messages or tag teen users.
In addition to privacy measures, sensitive content settings on teen accounts will be adjusted to the most restrictive options available. These default settings are designed to minimize teens’ exposure to inappropriate or harmful material.
Under these changes, teenagers can only be tagged or messaged by accounts they already follow or are connected to, further enhancing the safety of young users. This adds an additional layer of security by limiting unsolicited interactions on the platform.
Meta is not only focusing on teen privacy but also on parental involvement. Parents will now have access to a set of tools allowing them to monitor their children's interactions on Instagram and limit the amount of time spent on the app.
These parental controls include detailed monitoring features, enabling parents to review who their child is engaging with on the platform. Parents can also set specific restrictions on messaging and app usage, ensuring a safer digital environment for their children.
Teens under 16 will need parental permission to alter the default privacy settings on their accounts. This ensures that young users cannot independently make decisions that could expose them to risks without parental guidance.
The introduction of Instagram’s new privacy features comes in response to increasing concerns about the impact of social media on young users' mental health. Multiple studies have demonstrated links between social media use and heightened levels of anxiety, depression, and learning disabilities in teenagers.
Research has highlighted how excessive social media use, particularly by teenagers, can lead to mental health challenges such as depression and anxiety. These findings have fueled the push for stricter regulations on platforms like Instagram and TikTok.
Several reports show that prolonged exposure to social media can negatively affect adolescents, increasing the risk of learning disabilities and contributing to emotional and psychological struggles. As a result, there is growing concern about how young users interact with social platforms like Instagram.
Meta, along with other social media giants like ByteDance's TikTok and Google's YouTube, is currently facing legal action due to the addictive nature of their platforms. Several lawsuits have been filed on behalf of children and school districts, accusing these companies of failing to protect young users from harm.
In 2023, Meta and its competitors faced hundreds of lawsuits, including a significant legal challenge from 33 U.S. states, including California and New York. These states accuse Meta of misleading the public about the potential dangers of its platforms, especially regarding their impact on young users.
The lawsuits allege that platforms like Instagram contribute to addictive behaviors in children, further intensifying mental health risks. These legal challenges are part of a broader push for stricter regulation of social media companies.
In response to these concerns, the U.S. Senate has advanced two significant bills aimed at enhancing online safety for children and teens. The Kids Online Safety Act and The Children and Teens’ Online Privacy Protection Act would force social media companies to take responsibility for how their platforms affect younger users.
These bills, introduced in July 2024, aim to ensure that platforms like Instagram are more accountable for the safety and privacy of their younger audiences. If passed, social media companies will be required to implement stricter measures to safeguard young users from harmful content.
These legislative actions signal a significant shift in how social media platforms will be held accountable for their impact on children and teens. Companies like Meta will need to adopt stricter standards and provide tools that promote online safety.
Meta’s teen account rollout is being phased in over the next few months. The U.S., UK, Canada, and Australia will see these new privacy-focused accounts implemented within the next 60 days, while users in the European Union will receive them later this year.
The rollout will begin in key regions first, with other parts of the world following. Meta plans to introduce these new teen accounts globally starting in January 2025, ensuring a broader impact on young users worldwide.
Users in the U.S., UK, Canada, and Australia will be among the first to experience the new privacy settings. The move represents Meta’s commitment to addressing global concerns regarding online safety for teenagers.
Meta has announced that teenagers around the world will start to receive these teen accounts from January 2025, ensuring that this safety initiative reaches young Instagram users globally.