The threat of deepfakes

We need AI-backed technological tools to detect the unreal

January 20, 2021 12:15 am | Updated 12:15 am IST

Artificial Intelligence (AI) concept responsive landing page or web template with 3D isometric illustration of digital light ball between emerging rays.

Artificial Intelligence (AI) concept responsive landing page or web template with 3D isometric illustration of digital light ball between emerging rays.

The protesters who created chaos in Capitol Hill on January 6 believe that the 2020 U.S. election was stolen by the Democrats. This is largely due to misinformation and disinformation of which deepfakes are a part. Deepfakes — synthetic media, meaning media (including images, audio and video) that are either manipulated or wholly generated by Artificial Intelligence — even have the power to threaten the electoral outcome of the world’s oldest democracy. Sevreal social media platforms blocked President Donald Trump’s accounts after the attack.

Fabricating content

The cyberworld has been facing the challenge of deepfakes for a while now. AI is used for fabricating audios, videos and texts to show real people saying and doing things they never did, or creating new images and videos. These are done so convincingly that it is hard to detect what is fake and what is real. Detection can often be done only by AI-generated tools. Several books caution us against the threats of AI-generated content comprising non-existent personalities, synthetic datasets, unreal activities of real people, and content manipulation. Deepfakes can target anyone, anywhere. They are used to tarnish reputations, create mistrust, question facts, and spread propaganda.

In October 2020, the U.S. Senate summoned Facebook’s Mark Zuckerberg, Twitter’s Jack Dorsey and Google’s Sundar Pichai to find out what they are doing to tackle online misinformation, disinformation and fabricated content. Senators said they were worried about both censorship and the spread of misinformation. According to Section 230 of the Communications Decency Act of 1996, a law which protects freedom of expression and innovation on the Internet, “No provider or user of an interactive computer service shall be treated as the publisher or speaker of any information provided by another information content provider.” This means that the companies are not responsible for the posts on their platforms. The chief executives said they need the law to moderate content, but industry watchers and some politicians feel that the law is outdated and needs to be revisited.

India also faces the same problem. So far, it has not enacted any specific legislation to deal with deepfakes, though there are some provisions in the Indian Penal Code that criminalise certain forms of online/social media content manipulation. The Information Technology Act, 2000 covers certain cybercrimes. But this law and the Information Technology Intermediary Guidelines (Amendment) Rules, 2018 are inadequate to deal with content manipulation on digital platforms. (The guidelines stipulate that due diligence must be observed by the intermediate companies for removal of illegal content.) In 2018, the government proposed rules to curtail the misuse of social networks. Social media companies voluntarily agreed to take action to prevent violations during the 2019 general election. The Election Commission issued instructions on social media use during election campaigns. But reports show that social media platforms like WhatsApp were used as “vehicles for misinformation and propaganda” by major political parties during the election.

New tools

This is worrying. Existing laws are clearly inadequate to safeguard individuals and entities against deepfakes. Only AI-generated tools can be effective in detection. As innovation in deepfakes gets better, AI-based automated tools must be invented accordingly. Blockchains are robust against many security threats and can be used to digitally sign and affirm the validity of a video or document. Educating media users about the capabilities of AI algorithms could help.

In July 2020, the University of Washington and Microsoft convened a workshop with experts to discuss how to prevent deepfake technology from adversely affecting the 2020 U.S. presidential election. The workshop identified six themes: a) deepfakes must be contextualised within the broader framework of malicious manipulated media, computational propaganda and disinformation campaigns; b) deepfakes cause multidimensional issues which require a collaborative, multi-stakeholder response that require experts in every sector to find solutions; c) detecting deepfakes is hard; d) journalists need tools to scrutinise images, video and audio recordings for which they need training and resources; e) policymakers must understand how deepfakes can threaten polity, society, economy, culture, individuals and communities; and f) the idea that the mere existence of deepfakes causes enough distrust that any true evidence can be dismissed as fake is a major concern that needs to be addressed. In today’s world, disinformation comes in varied forms, so no single technology can resolve the problem. As deepfakes evolve, AI-backed technological tools to detect and prevent them must also evolve.

K. P. Shashidharan is a former director-general in the office of the Comptroller and Auditor General of India

0 / 0
Sign in to unlock member-only benefits!
  • Access 10 free stories every month
  • Save stories to read later
  • Access to comment on every story
  • Sign-up/manage your newsletter subscriptions with a single click
  • Get notified by email for early access to discounts & offers on our products
Sign in

Comments

Comments have to be in English, and in full sentences. They cannot be abusive or personal. Please abide by our community guidelines for posting your comments.

We have migrated to a new commenting platform. If you are already a registered user of The Hindu and logged in, you may continue to engage with our articles. If you do not have an account please register and login to post comments. Users can access their older comments by logging into their accounts on Vuukle.