top of page
Gen-AI Employee Support & Automation Platform

AI Chatbots Spreading Russian Propaganda


A new report from NewsGuard has revealed that several leading AI chatbots are unknowingly spreading Russian propaganda. This alarming discovery underscores the potential dangers of relying on AI for accurate information.



Why It Matters


As more users turn to AI chatbots for reliable answers, the revelation that these models can disseminate misinformation poses significant risks. This issue is particularly concerning as these AI tools are often marketed as providing accurate and trustworthy information.



Study Findings


NewsGuard's study involved entering 57 prompts into 10 of the most popular AI chatbots. The results were concerning:

- Misinformation Spread: The chatbots spread Russian disinformation 32% of the time.

- Source of Misinformation: Many narratives originated from John Mark Dougan, an American fugitive spreading misinformation from Moscow.

- Chatbots Tested: The tested AI models included OpenAI's ChatGPT-4, You.com, Grok, Inflection, Mistral, Microsoft's Copilot, Meta AI, Anthropic's Claude, Google Gemini, and Perplexity.



Responses and Reactions


Despite reaching out to the companies behind these AI models, NewsGuard has yet to receive a response. Steven Brill, co-CEO of NewsGuard, expressed deep concern over the findings. "The frequent repetition of well-known hoaxes and propaganda by these chatbots is alarming," Brill said. He urged users to be cautious about trusting AI-generated responses on controversial topics.



The Bigger Picture


The timing of this report is particularly critical as it comes in a year with significant global political events, including the U.S. presidential election. Using AI to spread misinformation could have profound implications for the democratic process.


Senator Mark Warner (D-Va.) highlighted the growing threat of misinformation in the current climate. "This is a real threat at a time when Americans are more willing to believe conspiracy theories than ever before," said Warner, who chairs the Senate Intelligence Committee.



Industry Response


Despite promises made at the Munich Security Conference to curb the spread of deepfakes and election-related misinformation, AI companies' actual efforts have been lacking. Warner expressed his disappointment, saying, "Where's the beef? I'm not seeing lots of activity."



NewsGuard Under Scrutiny


While NewsGuard continues its critical work, the House Oversight Committee is investigating it. Chairman James Comer has raised concerns about NewsGuard's potential role in censorship. NewsGuard has defended its actions, stating that its work with the Defense Department is focused on combating hostile disinformation from foreign governments.



Conclusion


The NewsGuard report serves as a wake-up call about AI chatbots' vulnerabilities in spreading misinformation. As technology advances, it is crucial to address these challenges to ensure the integrity of information in the digital age.

bottom of page