Facebook’s safety head tells U.K. lawmakers it does not amplify hate

Governments in Europe and the United States are grappling with regulating social media platforms to reduce the spread of harmful content, particularly for young users.

October 29, 2021 10:24 am | Updated 11:13 am IST

People are silhouetted as they pose with mobile devices in front of a screen projected with a Facebook logo, in this picture illustration.

People are silhouetted as they pose with mobile devices in front of a screen projected with a Facebook logo, in this picture illustration.

Facebook Inc's algorithms demote rather than promote polarising content, its global head of safety told British lawmakers on Thursday, adding that the U.S. company would welcome effective government regulation.

(Sign up to our Technology newsletter, Today's Cache, for insights on emerging themes at the intersection of technology, business and policy. Click here to subscribe for free.)

Governments in Europe and the United States are grappling with regulating social media platforms to reduce the spread of harmful content, particularly for young users.

Also Read | Facebook, Google, Twitter face grilling by U.K. lawmakers

Britain is leading the charge by bringing forward laws that could fine social media companies up to 10% of their turnover if they fail to remove or limit the spread of illegal content.

Secondary legislation that would make company directors liable could be proposed if the measures do not work.

Facebook whistleblower Frances Haugen told the same committee of lawmakers on Monday that Facebook's algorithms pushed extreme and divisive content to users.

Facebook's Antigone Davis denied the charge.

"I don't agree that we are amplifying hate," Davis told the committee on Thursday, adding: "I think we try to take in signals to ensure that we demote content that is divisive for example, or polarising."

Also Read | Facebook asks employees to preserve internal documents for legal inquiries

She said she could not guarantee a user would not be recommended hateful content, but Facebook was using AI to reduce its prevalence to 0.05%.

"We have zero interest in amplifying hate on our platform and creating a bad experience for people, they won't come back," she said. "Our advertisers won't let it happen either."

Davis said Facebook, which announced on Thursday it would rebrand as Meta , wanted regulators to contribute to making social media platforms safer, for example in research into eating disorders or body image.

Also Read | Facebook invests billions in metaverse efforts as ad business suffers

"Many of these are societal issues and we would like a regulator to play a role," she said, adding Facebook would welcome a regulator with "proportionate and effective enforcement powers".

"I think criminal liability for directors is a pretty serious step and I'm not sure we need it to take action."

0 / 0
Sign in to unlock member-only benefits!
  • Access 10 free stories every month
  • Save stories to read later
  • Access to comment on every story
  • Sign-up/manage your newsletter subscriptions with a single click
  • Get notified by email for early access to discounts & offers on our products
Sign in

Comments

Comments have to be in English, and in full sentences. They cannot be abusive or personal. Please abide by our community guidelines for posting your comments.

We have migrated to a new commenting platform. If you are already a registered user of The Hindu and logged in, you may continue to engage with our articles. If you do not have an account please register and login to post comments. Users can access their older comments by logging into their accounts on Vuukle.