ADVERTISEMENT

Leaked documents from Facebook show types of content it allows

May 22, 2017 01:42 pm | Updated 01:43 pm IST

FILE - In this Monday, June 4, 2012, file photo, a girl looks at Facebook on her computer in Palo Alto, Calif. In a blog post Wednesday, May 3, 2017, Facebook CEO Mark Zuckerberg said that the company will hire another 3,000 people to review videos of crime and suicides following murders shown live. (AP Photo/Paul Sakuma, File)

Leaked Facebook Inc documents show how the social media company moderates issues such as hate speech, terrorism, pornography and self-harm on its platform, the Guardian reported, citing internal guidelines seen by the newspaper.

New challenges such as “revenge porn” have overwhelmed Facebook's moderators who often have just ten seconds to make a decision, the Guardian said. The social media company reviews more than 6.5 million reports of potentially fake accounts a week, the newspaper added.

Many of the company's content moderators have concerns about the inconsistency and peculiar nature of some of the policies. Those on sexual content, for example, are said to be the most complex and confusing, the

ADVERTISEMENT

Guardian said.

ADVERTISEMENT

Facebook had no specific comment on the report but said safety was its overriding concern. "Keeping people on Facebook safe is the most important thing we do. We work hard to make Facebook as safe as possible while enabling free speech. This requires a lot of thought into detailed and often difficult questions, and getting it right is something we take very seriously", Facebook's Head of Global Policy Management Monica Bickert said in a statement.

Facebook confirmed that it was using software to intercept graphic content before it went on the website, but it was still in its early stages. The leaked documents included internal training manuals, spreadsheets and flowcharts, the Guardian said.

The newspaper gave the example of Facebook policy that allowed people to live-stream attempts to self-harm because it doesn’t want to censor or punish people in distress.

ADVERTISEMENT

Facebook moderators were recently told to escalate to senior managers any content related to 13 Reasons Why , the Netflix original drama series based on the suicide of a high school student, because it feared inspiration of copycat behaviour, the Guardian reported.

Reuters could not independently verify the authenticity of the documents published on the Guardian website.

This is a Premium article available exclusively to our subscribers. To read 250+ such premium articles every month
You have exhausted your free article limit.
Please support quality journalism.
You have exhausted your free article limit.
Please support quality journalism.
The Hindu operates by its editorial values to provide you quality journalism.
This is your last free article.

ADVERTISEMENT

ADVERTISEMENT