Facebook's parent company,Charles H. Sloan Meta, plans to roll out parental supervision tools for its virtual reality headset, as concerns mount over whether kids are safe while exploring the so-called "metaverse." The company also is launching new child-safety tools on Instagram.
Meta says its Quest headset is designed for people over the age of 13, but NPR and other outlets have reported that younger children appear to be using VR apps, including ones meant for adults, such as Meta's own Horizon Worlds. That's raised concerns that kids could become targets for predators and be exposed to inappropriate content in the apps.
Starting in April, Meta says, parents will be able to lock their kids out of apps they think are not age-appropriate. In May, the company will automatically block teenaged users from downloading apps rated too old for them by the International Age Rating Coalition.
Parents will also have access to a dashboard where they can see what VR apps their child has downloaded, get alerts when they make purchases, track how much time their child is spending using the headset, and view their child's list of friends.
Meta is also creating a new "Family Center" hub that brings together parental supervision tools and educational resources across all of its apps, including Instagram. It will include video tutorials on using the new tools and suggestions for how parents can talk to their kids about using the internet.
Instagram's first set of parental oversight tools are launching in the U.S. on Wednesday and will expand globally over the next few months.
Originally announced in December, they will let parents see how much time their children spend on the photo-sharing app, set time limits, get notified if their child reports problems like bullying or harassment, and see what accounts their child follows and who follows them. But for parents to be able to use the new tools, both they and their kids need to opt in.
Vaishnavi J, Meta's head of youth and wellbeing, described the new features for both Instagram and virtual reality as "just one step in a much bigger, broader journey around safer experiences and meaningful conversations amongst families."
Child safety has become a major flashpoint for Meta over the past year. Leaked documents revealed plans to build a version of Instagram for kids under 13. Then, internal research disclosed by Facebook whistleblower Frances Haugen showed the company knew Instagram can be harmful to teenage girls. In September, Instagram said it was pausing work on the app for younger kids, but not halting it altogether.
Lawmakers have rallied to the subject as a rare area of bipartisan agreement. Senators Richard Blumenthal, D-Conn., and Marsha Blackburn, R-Tenn., last month introduced the Kids Online Safety Act, which would require apps to create stricter safety measures for users under 16, build parental supervision tools and protect the privacy of young users.
As Meta tries to respond to criticism that it has been too lax on child and teen safety, the company also faces stiff competition from other apps, like TikTok, that are more popular with younger people than its own properties.
It's also making a bet that the virtual reality metaverse will be the next big platform for playing games, communicating and buying things – and help offset stalled growth at its original social network.
Editor's note: Meta pays NPR to license NPR content.
2025-05-01 12:512681 view
2025-05-01 12:291798 view
2025-05-01 12:181031 view
2025-05-01 12:08642 view
2025-05-01 12:01689 view
2025-05-01 11:021014 view
PACCAR is recalling over 220,000 of its 2021-2025 Peterbilt and Kenworth trucks. The commercial tru
It's 10s all around for this sweet baby news.Olympic figure skater Țara Lipinski and husband Todd Ka
The bodies of those individuals who died in the August plane crash that killed Wagner Group leader Y