AI supercharges disinformation and censorship, report warns

A woman looks at her mobile phone as she sits on a balcony near a man standing by a window at a residential compound in Wuhan, Hubei province, China, March 10, 2020

A woman looks at her mobile phone as she sits on a balcony near a man standing by a window at a residential compound in Wuhan, Hubei province, China, March 10, 2020. REUTERS

What’s the context?

Gains in generative AI technology make disinformation and censorship easier, threatening human rights, Freedom House report says

  • AI leads to increase in disinformation, surveillance
  • Generative tools pose threats to human rights, electoral process
  • When designed and deployed properly, AI can counter disinformation

Rapid advances in artificial intelligence are boosting online disinformation and enabling governments to increase censorship and surveillance in a growing threat to human rights, a U.S. non-profit said in a report published on Wednesday.

Global internet freedom declined for the 13th consecutive year, with China, Myanmar and Iran having the worst conditions of the 70 countries surveyed by the Freedom on the Net report, which highlighted the risks posed by easy access to generative AI technology.

AI allows governments to "enhance and refine online censorship" and amplify digital repression, making surveillance, and the creation and spread of disinformation faster, cheaper, and more effective, said the annual report by Freedom House.

A CCTV camera, installed on the boundary wall of a house is pictured at Vaikom in the Kottayam district of the southern state of Kerala, India November 23, 2017
Go DeeperIn India's surveillance hotspot, facial recognition taken to court
An Iranian woman living in Turkey reacts during a protest following the death of Mahsa Amini, outside the Iranian consulate in Istanbul, Turkey, September 21
Go DeeperMahsa Amini: facial recognition to hunt down hijab rebels in Iran
Traffic flows under the surveillance closed-circuit television camera (CCTV) system along Bakuli street in Kampala, Uganda August 14, 2019
Go DeeperThe rise of surveillance tech in Africa: What you need to know

"AI can be used to supercharge censorship, surveillance, and the creation and spread of disinformation," said Michael J. Abramowitz, president of Freedom House. "Advances in AI are amplifying a crisis for human rights online."

By some estimates, AI-generated content could soon account for 99% or more of all information on the internet, overwhelming content moderation systems that are already struggling to keep up with the deluge of misinformation, tech experts say.

Governments have been slow to respond, with few countries passing legislation for the ethical use of AI, while also justifying the use of AI-based surveillance technologies such as facial recognition on the grounds of security.

Generative AI-based tools were used in at least 16 countries to distort information on political or social issues over the period June 2022 to May 2023, the Freedom House report noted, adding that the figure is likely an undercount.

Meanwhile, in at least 22 countries, social media companies were required to use automated systems for content moderation to comply with censorship rules.

With at least 65 national-level elections taking place next year including in Indonesia, India and the United States, misinformation can have major repercussions, with deepfakes already popping up from New Zealand to Turkey.

"Generative AI offers sophistication and scale to spread misinformation on a level that was previously unimaginable - it is a force multiplier of misinformation," said Karen Rebelo, deputy editor at BOOM Live, a fact-checking organisation based in Mumbai.

While AI is a "military-grade weapon in the hands of bad actors," in India political parties and their proxies are the biggest spreaders of misinformation and disinformation, she said, and it is not in their interest to regulate AI.

While companies such as OpenAI and Google have imposed safeguards to reduce some overtly harmful uses of their AI-based chatbots, these can be easily breached, Freedom House said.

Even if deepfakes are quickly exposed, they can "undermine public trust in democratic processes, incentivise activists and journalists to self-censor, and drown out reliable and independent reporting," the report noted.

"AI-generated imagery ... can also entrench polarisation and other existing tensions. In extreme cases, it could galvanise violence against individuals or whole communities," it added.

For all its pitfalls, AI technology can be enormously beneficial, the report noted, so long as governments regulate its use and enact strong data privacy laws, while also requiring better misinformation-detection tools and safeguards for human rights.

"When designed and deployed safely and fairly, AI can help people evade authoritarian censorship, counter disinformation, and document human rights abuses," said Allie Funk, Freedom House's research director for technology and democracy.

For example, AI is being increasingly used in fact checking and to analyse satellite imagery, social media posts and images to flag human rights abuses in conflict zones.

(Reporting by Rina Chandran. Editing by Zoe Tabary)

Context is powered by the Thomson Reuters Foundation Newsroom.

Our Standards: Thomson Reuters Trust Principles


  • Content moderation
  • Internet shutdowns
  • Surveillance
  • Tech and inequality
  • Tech regulation
  • Social media
  • Data rights

Get our data & surveillance newsletter. Free. Every week.

By providing your email, you agree to our Privacy Policy.

Latest on Context