The Persuasive Power of AI: How Chatbots Can Shift Political Views – and What We Can Do about It
Artificial intelligence is rapidly becoming interwoven into the fabric of our daily lives,from answering simple questions to offering complex advice. But a recent study from the University of Washington, Stanford University, and ThatGameCompany reveals a concerning trend: even brief interactions with AI chatbots like ChatGPT can subtly - yet significantly – sway people’s political opinions. This research underscores the urgent need to understand the persuasive power of these systems and develop strategies to mitigate their influence.
The Experiment: A Subtle Shift in Outlook
The study, presented at the Association for Computational Linguistics, involved over 300 participants – a roughly even split of Republicans and Democrats. Researchers tasked participants with forming opinions on four relatively obscure policy topics: covenant marriage, unilateralism, the Lacey Act of 1900, and multifamily zoning. Before engaging with AI, participants self-reported their existing knowledge of these issues and their initial stances.The core of the experiment involved interacting with ChatGPT. Participants were divided into groups, some interacting with a chatbot subtly “primed” with a specific political leaning (either “radical right US Republican” or a neutral stance – a hidden instruction they weren’t aware of). The other group interacted with a neutrally programmed model. Participants averaged five interactions with the chatbots,discussing the policy topics and,in a second task,simulating budget allocation as a city mayor.
The results were striking. Nonetheless of their initial political affiliation, participants consistently shifted their views in the direction of the chatbot’s bias. The conservative-leaning chatbot steered conversations towards veterans and public safety, downplaying education and welfare. The liberal-leaning model did the opposite. Crucially, this shift occurred after just a handful of interactions.
Why Does This happen? The Power of Framing
The researchers, led by Jillian Fisher, a doctoral student at the University of Washington, attribute this phenomenon to the way AI chatbots frame information. “We certainly no that bias in media or in personal interactions can sway people,” Fisher explains. “And we’ve seen a lot of research showing that AI models are biased. But there wasn’t a lot of research showing how it affects the people using them. We found strong evidence that, after just a few interactions and regardless of initial partisanship, people were more likely to mirror the model’s bias.”
This isn’t about the chatbot directly telling users what to think. It’s about subtly shaping the conversation, highlighting certain aspects of an issue while downplaying others. This framing effect, a well-documented psychological phenomenon, can powerfully influence our perceptions and beliefs.
The Role of AI Literacy: A Potential Shield
Interestingly, the study revealed a key mitigating factor: prior knowledge. Participants who reported higher levels of understanding about AI were less susceptible to the chatbot’s persuasive tactics. This suggests that AI literacy – understanding how these systems work, their inherent biases, and their potential for manipulation – can act as a shield against undue influence.
“These models are biased from the get-go, and it’s super easy to make them more biased,” notes co-senior author Katharina Reinecke, a professor at the University of Washington. “That gives any creator so much power. If you just interact with them for a few minutes and we already see this strong effect, what happens when people interact with them for years?”
Implications and Future Research
This research has profound implications for the future of online discourse and democratic processes. As AI becomes increasingly integrated into our information ecosystem, the potential for manipulation grows exponentially.
The researchers are now focusing on several key areas:
Long-term Effects: Investigating the cumulative impact of prolonged interaction with biased AI models.
Expanding the Scope: Testing the findings with other large language models beyond ChatGPT.
Educational Interventions: Developing strategies to enhance AI literacy and empower users to critically evaluate information generated by these systems.
Mitigation Strategies: Exploring technical solutions to reduce bias in AI models themselves.
Protecting Yourself in the Age of AI
The goal isn’t to demonize AI, but to approach it with informed awareness.Here are some steps you can take to protect yourself from undue influence:
Be Skeptical: Don’t accept AI-generated information at face value. Always cross-reference with reputable sources.
Consider the Source: Be aware of the potential biases of the AI model you’re interacting with.* Question the Framing: Pay attention to how the AI chatbot









