Adopting Beliefs or Superficial Mimicry? Investigating Nuanced Ideological Manipulation of LLMs

Authors

  • Demetris Paschalides University of Cyprus, Computer Science Department, Nicosia, Cyprus
  • George Pallis University of Cyprus, Computer Science Department, Nicosia, Cyprus
  • Marios D. Dikaiakos University of Cyprus, Computer Science Department, Nicosia, Cyprus

DOI:

https://doi.org/10.1609/icwsm.v19i1.35885

Abstract

Large Language Models (LLMs) have transformed natural language processing, but concerns have emerged about their susceptibility to ideological manipulation, particularly in politically sensitive areas. Previous research has largely focused on LLM biases through a binary Left vs. Right framework, often using explicit ideological prompts and fine-tuning with political question-answering datasets. In this work, we move beyond this binary approach to explore the extent to which LLMs can be influenced across a nuanced spectrum of political ideologies, from Progressive-Left to Conservative-Right. We introduce a novel multi-task dataset designed to reflect diverse ideological positions through tasks such as ideological question-answering, statement ranking, manifesto cloze completion, and Congress bill comprehension. By fine-tuning three LLMs—Phi-2, Mistral, and Llama-3—on this dataset, we evaluate their capacity to adopt and express these nuanced ideologies. Our findings indicate that fine-tuning significantly enhances nuanced ideological alignment, while explicit prompts provide only minor refinements. This highlights the models' susceptibility to subtle ideological manipulation, suggesting a need for more robust safeguards to mitigate these risks.

Downloads

Published

2025-06-07

How to Cite

Paschalides, D., Pallis, G., & Dikaiakos, M. D. (2025). Adopting Beliefs or Superficial Mimicry? Investigating Nuanced Ideological Manipulation of LLMs. Proceedings of the International AAAI Conference on Web and Social Media, 19(1), 1510–1529. https://doi.org/10.1609/icwsm.v19i1.35885