Instagram and Facebook are adding more parental controls. Critics say they aren’t enough

Instagram Facebook
The Meta logo is seen at the Vivatech show in Paris, France, June 14, 2023.
AP Photo/Thibault Camus, File

By BARBARA ORTUTAY AP Technology Writer

Instagram and Facebook’s parent company Meta is adding some new parental supervision tools and privacy features to its platforms as social media companies face increasing scrutiny over their effects on teen mental health.

But many of the features require minors — and their parents — to opt in, raising questions about how effective the measures are.

Instagram will now send a notice to teens after they block someone, encouraging them to let their parents “supervise” their account. The idea is to grab kids’ attention when they might be more open to parental guidance.

If a teen opts in, the system will let parents set time limits, see who their kid follows or is followed by, and allows them to track how much time the minor spends on Instagram. It does not let parents see message content.

Instagram launched parental supervision tools last year to help families navigate the platform and find resources and guidance. A sticking point in the process is that kids need to sign up if they want parents to supervise their accounts. It’s not clear how many teen users have opted in and Meta has not disclosed any numbers.

By making the feature optional, Meta says it is trying to “balance teen safety and autonomy” as well as prompt conversations between parents and their children.

When families do opt in, supervision allows parents to see how many friends their child has in common with accounts the child follows or is followed by. If the child is followed by someone none of their friends follow, it could raise a red flag that the teen does not know the person in real life.

This, Meta says, “will help parents understand how well their teen knows these accounts, and help prompt offline conversations about those connections.”

Jim Steyer, the CEO and founder of Common Sense Media, called the news a “smoke screen.”

“None of these new features address the negative impact their business model is having on the well-being of kids, including their mental health. We need national privacy laws to protect kids,” Steyer said in a statement.

Meta is also adding parental supervision tools already available on Instagram and on virtual reality product to Messenger. The opt-in feature lets parents see how much time their child spends on the messaging service and information such as their contact lists and privacy settings — but not who they are chatting with.

Such features can be useful for families in which parents are already involved in their child’s online life and activities. Experts say that’s not the reality for many people.

Last month, U.S. Surgeon General Vivek Murthy warned that there is not enough evidence to show that social media is safe for children and teens and called on tech companies to take “immediate action to protect kids now.”

Murthy told The Associated Press that while he recognizes social media companies have taken some steps to make their platforms safer, those actions are not enough. For instance, while kids under 13 are technically banned from social media, many younger children access Instagram, TikTok and other apps by lying about their age, either with or without their parents’ permission.

Murthy also said it’s unfair to expect parents to manage what their children do with rapidly evolving technology that “fundamentally changes how their kids think about themselves, how they build friendships, how they experience the world — and technology, by the way, that prior generations never had to manage,”

“We’re putting all of that on the shoulders of parents, which is just simply not fair,” Murthy said. His office didn’t respond to a request for comment on Meta’s latest actions.