Facebook says it is testing more tools to help the social network prevent people from sharing content that exploits children.
One is a pop-up message that it plans to display across Facebook apps to people who use search terms linked to child exploration. It will detail the consequences of viewing such content and provide information on how to get help from offender diversion organizations.
The other measure is focused on the “non-malicious sharing” of “viral, meme child exploitative content.” People who share that material will see a safety alert about the harms it can cause.
The alert includes a warning that the content violates Facebook’s rules along with the legal ramifications of sharing such material. Facebook will remove the content and report it to the National Center for Missing and Exploited Children (NCMEC), and delete accounts that promote such material.
Elsewhere, Facebook has updated its child safety policies. According to global head of safety Antigone Davis, Facebook will take down “profiles, Pages, groups and Instagram accounts that are dedicated to sharing otherwise innocent images of children with captions, hashtags or comments containing inappropriate signs of affection or commentary about the children depicted in the image.”
While images or videos that people share may not break Facebook’s rules by themselves, Davis notes that “the accompanying text can help us better determine whether the content is sexualizing children and if the associated profile, Page, group or account should be removed.”
In addition, the company has updated its reporting menu in more areas of Facebook and Instagram. Users can select an option called “involves a child” under the Nudity and Sexual Activity section. Facebook says material reported in this way will take priority for its content reviewers. It has also adopted Google’s Content Safety API to detect when posts may contain child exploitation and prioritize them for reviewers.