For years, we’ve heard that AI chatbots are politically biased—skewing liberal, conservative, or somewhere in between. But a new study from researchers at the University of Klagenfurt suggests something surprising: most AI models aren’t as biased as we think—they just prefer not to engage in ideological debates at all.
By applying a statistical technique called Item Response Theory (IRT), the researchers found that large language models (LLMs) like ChatGPT 3.5 and Meta’s LLaMa don’t necessarily “lean” left or right. Instead, they often refuse to take a clear stance on political or economic issues. In other words, what looks like bias may actually be an avoidance strategy built into AI safety mechanisms.
The problem with existing bias detection methodsMost previous studies assessing bias in LLMs have taken one of two flawed approaches:
The researchers introduce an Item Response Theory (IRT)-based model, which is widely used in psychometrics and social science to assess latent traits—things that cannot be directly observed but can be inferred from responses to structured prompts.
The study applies two IRT models to LLMs:
To test whether LLMs exhibit bias, the researchers fine-tuned two families of models to explicitly represent left-leaning and right-leaning viewpoints:
These fine-tuned models served as baselines for bias assessment. Their responses were compared to off-the-shelf, non-fine-tuned models to see how ideological leanings manifested—or if they did at all.
Testing processOne of the study’s most striking findings is that off-the-shelf LLMs tend to avoid ideological questions rather than express a clear political bias. ChatGPT, for example, refused to answer 92.55% of ideological prompts, while the base LLaMa model avoided responding 55.02% of the time. This suggests that AI models are designed to lean toward neutrality or non-engagement rather than taking a partisan stance. Instead of actively skewing towards one political ideology, these models seem to default to avoiding controversial topics altogether, challenging previous claims of inherent bias in AI.
When examining fine-tuned models, the researchers found that expected ideological patterns did emerge—but only when the LLMs were specifically trained to adopt a political viewpoint. The fine-tuned “Left-GPT” and “Right-GPT” models produced predictable responses aligned with U.S. liberal and conservative ideologies. However, this bias did not appear in the non-fine-tuned versions, suggesting that ideological leanings in LLMs are not intrinsic but rather the result of intentional modifications during training.
The study also revealed that detecting bias in AI is more complex than simply categorizing responses as left-leaning or right-leaning. Some ideological test items were far more likely to trigger bias than others, highlighting the importance of issue selection in evaluating AI behavior. Economic issues, such as taxation and government spending, were particularly strong predictors of ideological bias compared to certain social issues. This indicates that not all political topics elicit the same level of response variation, making it crucial to assess how different types of prompts influence AI-generated outputs.
Gamification 2.0: How AI knows what keeps you engaged
Why this mattersThese findings challenge the prevailing assumption that LLMs inherently favor one political ideology over another. Instead, the evidence suggests that AI developers have prioritized non-engagement over taking a stance. While this may seem like a neutral approach, it raises new concerns about the way AI models interact with politically sensitive topics and the broader implications for AI governance, misinformation detection, and content moderation.
One key takeaway is that regulating AI bias is more complicated than previously thought. If AI models are systematically designed to avoid engagement, then efforts to ban “biased” AI outputs could inadvertently reinforce neutrality as the default position, leading to a lack of meaningful discourse on public policy, ethics, and governance. While neutrality may seem preferable to overt bias, it could also mean that AI-generated content sidesteps crucial discussions entirely, limiting its usefulness in politically charged conversations.
The study also underscores the need for more nuanced bias detection tools that differentiate between genuine ideological bias and response avoidance. Many previous studies may have misinterpreted non-engagement as an ideological stance, falsely labeling LLMs as partisan. Future bias detection methods should be designed to identify whether AI responses reflect a political position or whether they are simply programmed to steer clear of ideological engagement altogether.
Bias in AI is not just about what models say, but what they refuse to say. And that, perhaps, is the bigger story.
Featured image credit: Kerem Gülen/Midjourney