Deepfakes deceive voters in India, Pakistan before elections
• ‘Doctored’ videos circulating in India, Pakistan, Bangladesh and Indonesia
• Experts blame social media platforms, govts for not doing enough to curb their proliferation
• Meta, Google say they have policies in place to check deepfakes
Divyendra Singh Jadoun was busy making artificial intelligence-based visual effects and voice clones for film and television in India, when he began getting calls from politicians: could he create AI videos, or deepfakes, for their election campaign?
With a hotly contested local election in his home state of Rajasthan last November, and a national election due by May this year, the opportunity for his company, The Indian Deepfaker, is tremendous. But Jadoun was reluctant.
“The technology to create deepfakes is so good now, it can be done almost instantaneously, with very little effort — and people cannot tell if it’s real or fake,” said Jadoun. “There are no guidelines on deepfakes, and that’s worrying, as it has the potential to influence how a person votes,” he said.
Instagram reels of Indian Prime Minister Narendra Modi singing in regional languages have gone viral recently, as have TikTok videos of Indonesian presidential candidates Prabowo Subianto and Anies Baswedan speaking in fluent Arabic.
But they were all created with AI, and posted with no label.
With elections due in India, Indonesia, Bangladesh and Pakistan in the coming weeks, misinformation is rife on social media platforms, with deepfakes — video or audio made using AI and broadcast as authentic — being particularly concerning, say tech experts and authorities.
In India, where more than 900m people are eligible to vote, Modi has said deepfake videos are a “big concern”, and authorities have warned social media platforms they could lose their safe-harbour status that protects them from liability for third-party content posted on their sites if they do not act.
In Pakistan, where an election is scheduled for Feb 8, former premier Imran Khan, who is in prison in an Official Secrets Acts case, used an AI-generated image and voice clone to address an online election rally in December, which drew more than 1.4 million views on YouTube and was attended live by tens of thousands.
While Pakistan has drafted an AI law, digital rights activists have criticised the lack of guardrails against disinformation, and to protect vulnerable communities including women.
“The threat that disinformation poses to elections and the overall democratic process in Pakistan cannot be stressed upon enough,” said Nighat Dad, co-founder of the non-profit Digital Rights Foundation.
“In the past, disinformation on online platforms has managed to sway voting behaviour, party support, and even influenced legislation change. Synthetic media will make this easier to do,” she added.
In Indonesia, where over 200m voters will go to polls on Feb 14, deepfakes of all three presidential candidates and their running mates are circulating online, and have the potential to influence election outcomes, said Nuurrianti Jalli, who studies misinformation on social media.
“From microtargeting of voters with disinformation to spreading false narratives at a scale and speed unachievable by human actors alone, these AI tools can significantly influence voter perceptions and behaviour,” she said.
“In environments where misinformation is already prevalent, AI-generated content can further skew public perception and influence voting behaviour,” added Jalli, an assistant professor at Oklahoma State University’s media school.
Political propaganda
Deepfake images and videos churned out by generative AI tools such as Midjourney, Stable Diffusion and OpenAI’s Dall-E popped up ahead of elections from New Zealand to Turkey and Argentina last year, with growing concerns about their impact on U.S. presidential polls in November.
AI makes the creation and spread of disinformation faster, cheaper and more effective, the US non-profit Freedom House said in a recent report.
In Bangladesh, where Prime Minister Sheikh Hasina is set for her fourth straight term after polls on Jan 7, deepfake videos of female opposition politicians Rumin Farhana in a bikini and Nipun Roy in a swimming pool have emerged.
While they were debunked quickly, they are still circulated, and even poor-quality deepfake content is misleading people, said Sayeed Al-Zaman, an assistant professor of journalism at Bangladesh’s Jahangirnagar University, who studies social media.
“Given the low levels of information and digital literacy in Bangladesh, deepfakes can be potent carriers of political propaganda if crafted and deployed effectively,” he said.
“But the government does not appear concerned.”
The ministry of information did not respond to a request for comment.
Dangerous sign
At least 500,000 video and voice deepfakes were shared on social media sites globally in 2023, estimated DeepMedia, a company developing tools to detect synthetic media.
Platforms have struggled to keep up.
Meta, which owns Facebook, Instagram and WhatsApp, said it aims to remove synthetic media when the “manipulation is not apparent and could mislead, particularly in the case of video content.”
Google, which owns YouTube, said in November that the video sharing platform requires “creators to disclose altered or synthetic content that is realistic, including using AI tools, and we’ll inform viewers about such content through labels”.
But countries including India, Indonesia and Bangladesh have recently passed laws to more closely police online content and penalise social media sites for content deemed misinformation, so platforms are “holding their punches”, said Raman Jit Singh Chima, Asia policy director at advocacy group, Access Now.
In these countries, “this election cycle is actually worse than the last cycle — platforms are not set up to handle problems, and they are not being responsive and proactive enough. And that’s a very dangerous sign,” he said.
“There is a danger that the world’s attention is only on the U.S. election, but the standards being applied there, the effort being made there should be duplicated everywhere,” he added.
Published in Dawn, January 4th, 2024