The Santa Clara Principles

On Transparency and Accountability in Content Moderation


Top

Facebook’s Response

Thank you for your November 13 letter to Mark Zuckerberg addressing notice, appeal, and data transparency for violations of Facebook’s Community Standards. You’ve raised important questions about how Facebook is approaching these issues.

Your letter gives us an opportunity to summarize the work we’ve been doing over the past year in these areas. Please find details below, using the headings in your letter. We have also noted areas where we aren’t currently in line with your recommendations, or where we are working in that direction.

Please bear in mind that much of this is work in progress, and we will provide further updates as our policies and enforcement develop. We look forward to continuing this dialog with you.

Notice

Our procedures for notifying users of Community Standards violations have become more detailed over time. In April, we released the internal guidelines https://newsroom.fb.com/news/2018/04/comprehensive-community-standards/that we use to enforce on our Community Standards, so that people have clarity on exactly where we draw the line. As part of this process, we have been working to improve the notifications that we send users when we remove content that goes against our Community Standards.

In the majority of cases, when we tell people their content goes against our standards, we cite the policy violated. In a few cases, where concerns for safety or gaming of our policies are high, such as with our policies around sexual exploitation and terrorist propaganda, we provide general notice that the content in question has violated our Community Standards. We also identify for users the specific piece of content that violates our standards.

With regard to how content is detected, evaluated, and removed (if it violates our policies), two points:

First, Facebook’s Community Standards Enforcement Report https://transparency.facebook.com/community-standards-enforcementcontains numbers showing how much violating content we have detected on our service. It’s an important part of our effort to be transparent — so people can judge for themselves how well we are doing. As outlined further below, this report provides aggregate data for most major policies on the percentage of violating content that we have proactively detected as compared to that which was reported by users. For six of the eight policies for which we report data in the enforcement report, we identified over 95% of the violating content ourselves. This figure is significant because it indicates that we are successfully removing large volumes of violating content before we receive user reports. User reports nonetheless remain important, not least because they provide a signal that users perceive the reported content as harmful.

Second, when it comes to individual pieces of content, we don’t disclose how we are made aware of violating content, since doing so could undermine the confidentiality of user reporting — an important principle underlying our enforcement approach.

Appeals

When we talk about appeals, we’re referring to ways for people to request re-review of a content decision we’ve made. Prior to this year, such re-review was available to people whose profiles, Pages, or Groups had been taken down. In April 2018, we introduced the option to request re-review of individual pieces of content that were removed for adult nudity or sexual activity, hate speech, or graphic violence. We’ve subsequently extended this option so that re-review is now available for additional content areas, including dangerous organizations and individuals (which includes our policies on terrorist propaganda), bullying and harassment, regulated goods, and spam. We are continuing to roll out re-review for additional types of violations. We also plan to launch this option for individuals who have reported content that was not removed.

When we announced the ability to seek re-review of content removals, we also said that we would like to set up a system where users can provide more information and context on decisions that they think we got wrong. Along these lines, we are currently experimenting with the best ways to solicit context from users.

There are some violation types – for example, severe safety policy violations – for which we don’t offer re-review. For all other types of content, in order to request re-review of a content decision we’ve made, you are given the option to “Request Review.” We make the opportunity to request this review clear, either via a notification or interstitial. Re-review is conducted by a different human reviewer than that of the original review. In a limited number of cases, where we have very high confidence in our original decision and believe that human review is not efficient — for example, spam and cases involving “banked” content — we may also rely on automation for re-review. If we find we have made a mistake, we will restore the content. Typically, the re-review takes place within 24 hours.

Please take a look at the three screenshots appearing at the bottom of this message. These images show, respectively, notice of a Community Standards violation; the option to seek re-review of our content decision; and acknowledgment of a request for re-review.

We welcome the opportunity for Facebook to collaborate with stakeholders on innovative approaches to self-regulation and accountability. We engage regularly with external groups and experts on our policy development and enforcement, and we are looking to do more. For example, our CEO recently laid out a path for appealing content decisions to an independent body https://www.facebook.com/notes/mark-zuckerberg/a-blueprint-for-content-governance-and-enforcement/10156443129621634/(see “Independent Governance and Oversight”). In 2019, we’ll undertake a process of stakeholder consultation on this idea, to bring external voices into the decision making process. We would like to include your voices in that process.

Data

Our first ever Community Standards Enforcement Report, released in May 2018 as part of our larger bi-annual Transparency Report, provided data on the volume of content we actioned (including both content removed and content in which an interstitial or warning screen was added) and the percentage of violations we found before users reported them in the following areas: adult nudity and sexual activity, fake accounts, hate speech, spam, terrorist propaganda, and violent and graphic content.

We received your letter just two days before the release of our second Community Standards Enforcement Report https://transparency.facebook.com/community-standards-enforcement. This report contains a range of information on topics raised in the letter, covering the period from April to September 2018. The report includes updates to the data we shared in our May report, as well as new data in the areas of bullying and harassment, and child nudity and sexual exploitation of children. For each of these policies, we provided data on how much content we took action on, and the percentage of violations that we found before users reported them. Where possible, we also provided information on prevalence of this content on Facebook. The report further highlights data we plan to provide in the future, including an indication of how quickly we took action on Community Standards violations and how often we restore content upon re-review.

In many policy areas, the percentage of removed content first identified by Facebook’s automated systems is above 95%. In the areas of hate speech and bullying and harassment, where context is critical, proactive detection rates are lower (~52% and ~15% respectively). At this point, we do not provide a further breakdown of reporting source.

We also don’t provide data on the format of the content being removed, whether text, photo, or video, because we do not believe this data provides critical information to our users or civil society about our content review practices.

We audit the quality and accuracy of reviewer decisions on an ongoing basis, to understand where our policies or training can be improved, and to follow up with reviewers on improving where errors are being made. Reducing these errors is one of our most important priorities.

We hope this response is useful, and appreciate the opportunity to convey these updates. We are exploring ways to make our updates more easily accessible. In the meantime, please feel free to share this message with your colleagues and other members of civil society.