Facebook makes its community guidelines public and introduces an appeals process

April 24, 2018

Last May, The Guardian published a leaked copy of Facebook’s content moderation guidelines, which describe the company’s policies for determining whether posts should be removed from the service. Almost a year later, Facebook is making an expanded set of those guidelines available to the public, a move designed to gather input from users around the world. The company is also introducing a new appeals process, allowing users to request a review if they believe their post has been removed unfairly.

The community standards run 27 pages and cover topics including bullying, violent threats, self-harm, and nudity, among many other topics. “These are issues in the real world,” said Monika Bickert, head of global policy management at Facebook, in an interview with reporters. “The community we have using Facebook and other large social media mirrors the community we have in the real world. So we’re realistic about that. The vast majority of people who come to Facebook come for very good reasons. But we know there will always be people who will try to post abusive content or engage in abusive behavior. This is our way of saying these things are not tolerated. Report them to us, and we’ll remove them.”

The guidelines apply to every country in which Facebook operates, and have been translated into more than 40 languages. The company says it developed them in conjunction with a “couple hundred” of experts and advocacy groups representing the entire world. As the guidelines evolve—and they will evolve, Bickert said—they will be updated simultaneously in every language.

The guidelines mostly apply to other Facebook services, including Instagram, although there are differences. (You don’t have to use your real name on Instagram, for one.) The underlying policies haven’t changed, Bickert said, though they now include extra guidance on making decisions. “What’s changing is the level of explanation about how we apply those policies,” Bickert said.

Amid a series of unfolding humanitarian crises, Facebook has been under pressure to improve content moderation around the globe. In March, the United Nations blamed Facebook for spreading hatred of the Rohingya minority. Facebook was also forced to temporarily shut down its services in Sri Lanka last month after inflammatory messages posted to the service incited mob violence against the country’s Muslim minority. This weekend, a report in The New York Times connected hate speech on Facebook to murders in Indonesia, India, and Mexico.

In response, the company has said it will double its 10,000-person safety and security team by the end of this year. It also plans to update the guidelines regularly as new threats emerge. Facebook is making the guidelines public now because it hopes to learn from users’ feedback, Bickert said.

Facebook also announced plans to develop a more robust process for appealing takedowns that were made in error. The company has faced regular criticism for high-profile takedowns over the years, whether it’s over a picture of a woman breastfeeding her child or an iconic wartime photo.

Now users will be able to request that the company review takedowns of content they posted personally. If your post is taken down, you’ll be notified on Facebook with an option to “request review.” Facebook will review your request within 24 hours, it says, and if it decides it has made a mistake, it will restore the post and notify you. By the end of this year, if you have reported a post but been told it does not violate the community standards, you’ll be able to request a review for that as well.

The Verge has the full story

Sponsored Recommendations

ASK THE EXPERT: ServiceNow’s Erin Smithouser on what C-suite healthcare executives need to know about artificial intelligence

Generative artificial intelligence, also known as GenAI, learns from vast amounts of existing data and large language models to help healthcare organizations improve hospital ...

TEST: Ask the Expert: Is Your Patients' Understanding Putting You at Risk?

Effective health literacy in healthcare is essential for ensuring informed consent, reducing medical malpractice risks, and enhancing patient-provider communication. Unfortunately...

From Strategy to Action: The Power of Enterprise Value-Based Care

Ever wonder why your meticulously planned value-based care model hasn't moved beyond the concept stage? You're not alone! Transition from theory to practice with enterprise value...

State of the Market: Transforming Healthcare; Strategies for Building a Resilient and Adaptive Workforce

The U.S. healthcare system is facing critical challenges, including workforce shortages, high turnover, and regulatory pressures. This guide highlights the vital role of technology...