As artificial intelligence (AI) systems increasingly influence everything from education to politics, a new study has found that many large language models (LLMs) exhibit a consistent left-wing political bias.
The research raises concerns about the role of AI in shaping public opinion and calls for greater transparency and oversight in the development of these systems.
“When probed with politically charged questions/statements, most conversational LLMs tend to generate responses that are diagnosed by most political testing instruments as exhibiting preferences for center-left views “wrote study author Dr. David Rozado. “As LLMs begin to partially replace traditional information sources like search engines and Wikipedia, the societal implications of the political biases embedded in LLMs are substantial.”
In an era where artificial intelligence (AI) is increasingly integrated into daily life, concerns about the neutrality of these systems take center stage.
As AI continues to evolve, its applications go beyond simple tools for convenience and productivity. Large language models, designed to mimic human conversation, are now used for tasks such as writing articles, answering complex questions and even providing mental health support.
With their broad reach and growing use in fields such as journalism, education, and customer service, these models are able to shape public discourse in unprecedented ways.
However, as the capabilities of these systems increase, so does the risk of unintended consequences. Political bias in LLMs could lead to presenting biased information to users, subtly guiding their thoughts on current issues such as the economy, social policies, and government.
Last year, Elon MuskCEO of SpaceX and X (formerly Twitter), launched Grok, a broad language model designed to compete with what he perceives as political bias in existing AI systems.
Musk has long spoken out about the risks of AI influencing public discourse, and Grok is part of his broader strategy to ensure AI does not unduly influence political views and , in his speech, own words“Stop the waking mind virus.”
A study published in PLOS ONE suggests that some of Musk’s concerns are valid. He reports that large language models (LLMs) such as GPT-4, Claude, and Llama 2, among others, often display political biases, tending toward left-wing ideologies.
These models, which underpin popular AI tools like ChatGPT, have the potential to influence societal perspectives and public discourse, sparking a growing conversation about the ethical implications of AI bias.
The study, led by Dr David Rozado, associate professor of computational social sciences at Otago Polytechnic in New Zealand, analyzed 24 conversational LLMs across a series of 11 political orientation tests. He concluded that most of these models consistently generated responses aligned with center-left political views.
This finding is particularly important as LLMs increasingly replace traditional information sources such as search engines, social media and academic resources, amplifying their influence on political opinions and worldviews of individual users.
Given that millions of people rely on LLMs to answer their questions and form opinions, the discovery of political leanings within these models raises ethical concerns that urgently need to be addressed.
Dr. Rozado’s study constitutes one of the most comprehensive analyzes of policy preferences embedded in LLMs. The research involved administering various political tests, including the widely used Political Compass Test and the Eysenck Political Test, to models such as GPT-3.5, GPT-4, Google’s Gemini, and Anthropic’s Claude.
During these tests, results showed that most models consistently provided answers classified as left-leaning on economic and social topics.
For example, in the political compass test, LLMs leaned primarily toward progressive ideals, such as social reform and government intervention, while downplaying more conservative perspectives emphasizing individual liberty and limited government.
Interestingly, the study also found significant variability between models, with some LLMs exhibiting more pronounced biases than others. Open source models like Llama 2 have been found to be slightly less biased than their closed counterparts, raising questions about the role of corporate control and proprietary algorithms in shaping AI bias.
The political leanings of large language models arise from several factors, many of which are deeply rooted in the data on which they are trained. LLMs are typically trained on large datasets compiled from publicly available sources, such as websites, books, and social media.
This data often reflects societal biases, which are reflected in AI models. Additionally, how these models are refined after their initial training can significantly influence their policy direction.
Dr. Rozado’s study go further to explore how political alignment can be intentionally built into AI systems through a process called Supervised Fine-Tuning (SFT). Researchers can bias models toward specific policy preferences by exposing LLMs to modest amounts of politically aligned data.
This finding is both a warning and an opportunity: while AI can be tailored to specific applications, that same capability can introduce biases that may not be immediately apparent to users.
“With modest calculation and politically personalized training data, a practitioner can align LLM policy preferences with target regions of the political spectrum via supervised fine-tuning,” Dr. Rozado wrote. “This provides evidence for the potential role of supervised fine-tuning in the emergence of policy preferences within LLMs.”
However, Dr. Rozado cautions that the results of his study should not be interpreted as evidence that organizations are deliberately injecting left-wing political biases into large language models (LLMs). Rather, it suggests that any consistent policy trends may result unintentionally from the instructions provided to annotators or from dominant cultural norms during the training process.
Although not explicitly political, these influences can shape LLM outcomes on a range of political topics due to broader cultural patterns and analogies in the semantic understanding of the patterns.
The discovery of political bias in LLMs is coming at once while trust in AI systems is already the subject of intense debate. As these models play an increasingly important role in shaping public discourse, the potential for them to unintentionally promote specific political ideologies is concerning.
Furthermore, as LLMs are adopted in fields such as education, journalism and law, their influence could have far-reaching consequences for democratic processes and public opinion.
The study’s findings highlight the need for transparency and accountability in AI development. Like these technologies continue to evolve, there is an urgent need for clear guidelines on how models are trained, what data they are exposed to and how they are refined. Without such measures, AI risks becoming a tool for reinforcing existing biases or, worse, subtly manipulating public opinion.
Experts say that as AI systems like LLMs become more and more integrated into the fabric of modern life, it is crucial that we address the ethical challenges posed by their use. Policymakers, developers, and the general public must demand greater transparency in how these models are constructed and ensure that they do not inadvertently shape political discourse in biased ways.
One potential solution is to introduce regular audits and checks to ensure that LLMs maintain their political neutrality or reveal any inherent bias. In addition, efforts to diversify The training data used to build these models could help reduce the risk of bias, ensuring a wider range of perspectives are represented.
Ultimately, as AI continues to shape the way we live, work, and interact with the world, it is crucial that these systems are designed with fairness and transparency in mind.
“Traditionally, people have relied on search engines or platforms like Wikipedia to quickly and reliably access a mix of factual and biased information. However, as LLMs become more advanced and accessible, they are beginning to partially replace these conventional sources,” concludes Dr. Rozado. “This change in the source of information has profound societal implications, as LLMs can shape public opinion, influence voting behavior, and impact the overall discourse of society. »
“Therefore, it is crucial to critically examine and address potential political biases embedded in LLMs to ensure balanced, fair and accurate representation of information in their responses to user queries. »
Tim McMillan is a retired law enforcement official, investigative journalist and co-founder of The Debrief. His writings generally focus on defense, national security, the intelligence community, and psychology-related topics. You can follow Tim on Twitter: @LtTimMcMillan. Tim can be reached by email: tim@thedebrief.org or by encrypted email: LtTimMcMillan@protonmail.com