Facebook’s safety head tells U.K. lawmakers it does not amplify hate

People are silhouetted as they pose with mobile devices in front of a screen projected with a Facebook logo, in this picture illustration.   | Photo Credit: Reuters

Facebook Inc's algorithms demote rather than promote polarising content, its global head of safety told British lawmakers on Thursday, adding that the U.S. company would welcome effective government regulation.

(Sign up to our Technology newsletter, Today's Cache, for insights on emerging themes at the intersection of technology, business and policy. Click here to subscribe for free.)

Governments in Europe and the United States are grappling with regulating social media platforms to reduce the spread of harmful content, particularly for young users.

Also Read | Facebook, Google, Twitter face grilling by U.K. lawmakers

Britain is leading the charge by bringing forward laws that could fine social media companies up to 10% of their turnover if they fail to remove or limit the spread of illegal content.

Secondary legislation that would make company directors liable could be proposed if the measures do not work.

Facebook whistleblower Frances Haugen told the same committee of lawmakers on Monday that Facebook's algorithms pushed extreme and divisive content to users.

Facebook's Antigone Davis denied the charge.

"I don't agree that we are amplifying hate," Davis told the committee on Thursday, adding: "I think we try to take in signals to ensure that we demote content that is divisive for example, or polarising."

Also Read | Facebook asks employees to preserve internal documents for legal inquiries

She said she could not guarantee a user would not be recommended hateful content, but Facebook was using AI to reduce its prevalence to 0.05%.

"We have zero interest in amplifying hate on our platform and creating a bad experience for people, they won't come back," she said. "Our advertisers won't let it happen either."

Davis said Facebook, which announced on Thursday it would rebrand as Meta, wanted regulators to contribute to making social media platforms safer, for example in research into eating disorders or body image.

Also Read | Facebook invests billions in metaverse efforts as ad business suffers

"Many of these are societal issues and we would like a regulator to play a role," she said, adding Facebook would welcome a regulator with "proportionate and effective enforcement powers".

"I think criminal liability for directors is a pretty serious step and I'm not sure we need it to take action."

Our code of editorial values

Related Topics
This article is closed for comments.
Please Email the Editor

Printable version | Jan 19, 2022 2:18:44 AM |

Next Story