Meta

Community Standards Enforcement Report, First Quarter 2022

By Guy Rosen, VP of Integrity

Takeaways

  • The prevalence of harmful content on Facebook and Instagram remained relatively consistent — and decreased in some of our policy areas mostly due to improving and expanding our proactive detection technology.
  • We’re releasing the findings from EY’s independent assessment of our 2021 fourth quarter report to make sure we’re calculating and reporting on our metrics accurately.

Today we’re publishing our Community Standards Enforcement Report for the first quarter of 2022. It shows how we enforced our policies from January through March of 2022 across 14 policy areas on Facebook and 12 on Instagram. We’re also sharing several other reports in our Transparency Center including: 

We’re also releasing the findings from EY’s independent assessment of our enforcement reporting. Last year, we asked EY to verify the metrics of our Community Standards Enforcement Report since we don’t believe we should grade our own homework. EY’s assessment concluded that the calculation of the metrics have been prepared based on the specified criteria and the metrics are fairly stated, in all material respects. As we keep growing this report, we will also keep working on ways to make sure that it’s accurately presented and independently verified.

Highlights From the Report

The prevalence of violating content on Facebook and Instagram remained relatively consistent but decreased in some of our policy areas from Q4 2021 to Q1 2022. 

On Facebook in Q1 we took action on:

  • 1.8 billion pieces of spam content, which was an increase from 1.2 billion in Q4 2021, due actions on a small number of users making a large volume of violating posts
  • 21.7 million pieces of violence and incitement content, which was an increase from 12.4 million in Q4 2021, due to the improvement and expansion of our proactive detection technology.

On Instagram in Q1 we took action on:

  • 1.8 million pieces of drug content, which was an increase from 1.2 million from Q4 2021, due to updates made to our proactive detection technologies.

We also saw an increase in the proactive detection rate of bullying and harassment content from 58.8% in Q4 2021 to 67% in Q1 2022 thanks to improvement and expansion of our proactive detection technology. We also continued to see a slight decrease in prevalence on Facebook from 0.11-0.12% in Q4 2021 to 0.09% in Q1 2022.

Graphic showing bullying and harassment prevalence during Q1 2022

Refining Our Policies and Enforcement

Over the years we’ve invested in building technology to improve how we can detect violating content. With this progress we’ve known that we’ll make mistakes, so it’s been equally important along the way to also invest in refining our policies, our enforcement and the tools we give to users.

For example, we’ve improved our transparency over the years to better inform people why we took down a post and we’ve improved the ability to appeal and ask us to take another look. We include metrics about appeals in this report. Recently we’ve begun to evaluate the effectiveness of our penalty system more deeply, for example testing the impact of giving people additional warnings before triggering more severe penalties for violating our policies.

Last year we saw how policy refinements can ensure that we aren’t over enforcing beyond what we intend. Updates we made to our bullying and harassment policy better accounted for language that can be easily misunderstood without context. Enforcement systems also play a role: we recently began testing new AI technology that identifies and prevents potential over-enforcement by better learning from content that’s appealed and subsequently restored. For example, the same word that is an offensive slur in the U.S. is also a common British term for cigarette which would not violate our policies. We’re also testing a number of improvements to our proactive enforcement, to enable some admins of Groups to better shape group culture and take context into account around what is and isn’t allowed in their space; or to better reflect the context in which people write comments between friends, where sometimes good-natured banter could be mistaken as violating content.

Looking Ahead

For a long time, our work has focused on measuring and reducing the prevalence of violating content on our services. But we’ve worked just as hard to improve the accuracy of our enforcement decisions. While prevalence helps us measure what we miss, we’ve also been working on developing robust measurements around mistakes, which will help us better understand where we act on content incorrectly. We believe sharing metrics around both prevalence and mistakes will provide a more complete picture of our overall enforcement system and help us improve, so we are committed to providing this in the future.

Finally, as new regulations continue to roll out around the globe, we are focused on the obligations they create for us. So we are adding and refining processes and oversight across many areas of our work. This will enable us to make continued progress on social issues while also meeting our regulatory obligations more effectively.



To help personalize content, tailor and measure ads, and provide a safer experience, we use cookies. By clicking or navigating the site, you agree to allow our collection of information on and off Facebook through cookies. Learn more, including about available controls: Cookie Policy