We tested 20 LLMs for ideological bias, revealing distinct alignments

Methodology & Limitations

  • Commenters find the prompt set narrow: few questions per axis, English-only, forced A/B/pass answers, and a single system prompt.
  • Suggestions: expand question bank, add multiple phrasings per item, randomize option order, and translate prompts to see how different language “Overton windows” shift answers.
  • Small sample size per axis means a model can flip from 0% to 100% in a category just by changing 1–2 answers, so per-axis claims are seen as fragile.

Nature and Sources of Bias

  • Many argue bias is unavoidable: models inherit it from human data; selection of training sources (news, publishers, social platforms) skews toward elite or “progressive” online discourse rather than general populations.
  • Others stress bias isn’t layered on a neutral engine; it’s baked into the weights through data and RLHF.
  • Labels like “progressive” and “conservative” are criticized as narrow (heavily tied to abortion/gender/social-norm questions) and not capturing economic or geopolitical stances; several see most models as “establishment / neoliberal” rather than truly progressive.

Interpretation of Results

  • Observed pattern: almost all models lean regulatory and socially progressive, with a few outliers leaning libertarian or conservative on some axes. Some call this polarization; others call it uniformity.
  • Specific findings spark debate: pro-UN/NATO/EU stances, pro-Israel answers, disagreement with international law on resistance to occupation.
  • Distillation and newer versions sometimes show ideological shifts, raising questions about what alignment steps changed.

Control, Neutrality, and “Truth”

  • One thread: are models mostly shaped by data, or are companies intentionally steering politics? Both possibilities are considered.
  • Several argue it’s impossible to define a single “neutral” position; “unbiased” might mean mirroring empirical consensus, not 50/50 on every controversy.
  • Others propose LLMs should refuse direct answers on hot-button questions, instead outlining competing arguments—but even choosing which “facts” to present is itself contested.

Social & Political Implications

  • Concerns that LLM-backed devices could become new chokepoints: effectively banning or monetizing disfavored activities, similar to payment-processor control.
  • Fear that ideologues will game training data and alignment to capture “perceived truth,” using LLMs as a new propaganda vector.
  • Counterview: LLMs are tools; healthy users should treat them as maps of discourse, not oracles.

Proposed Responses

  • Add explicit political-bias sections to model cards; place models on ideological compasses, including compliance rates.
  • Provide transparency about training and alignment and allow user choice among differently biased models, analogous to picking newspapers or feed algorithms.
  • Accept that the real task is not “removing” bias but choosing which bias is appropriate for a given application.