State media control influences large language models

  • Palmer, A. & Spirling, A. Large language models can argue in convincing ways about politics, but humans dislike AI authors: implications for governance. Polit. Sci. 75, 281–291 (2023).

    Article 

    Google Scholar
     

  • Bai, H. et al. LLM-generated messages can persuade humans on policy issues. Nat. Commun. 16, 6037 (2025).

    Article 
    ADS 
    PubMed 
    PubMed Central 

    Google Scholar
     

  • Hackenburg, K. & Margetts, H. Evaluating the persuasive influence of political microtargeting with large language models. Proc. Natl Acad. Sci. USA 121, e2403116121 (2024).

    Article 
    PubMed 
    PubMed Central 

    Google Scholar
     

  • Salvi, F. et al. On the conversational persuasiveness of GPT-4. Nat. Hum. Behav. 9, 1645–1653 (2025).

    Article 
    PubMed 
    PubMed Central 

    Google Scholar
     

  • Costello, T. H., Pennycook, G. & Rand, D. G. Durably reducing conspiracy beliefs through dialogues with AI. Science 385, eadq1814 (2024).

    Article 
    ADS 
    CAS 
    PubMed 

    Google Scholar
     

  • Carrasco-Farre, C. Large language models are as persuasive as humans, but how? About the cognitive effort and moral-emotional language of LLM arguments. Preprint at https://arxiv.org/abs/2404.09329 (2024).

  • Tessler, M. H. et al. AI can help humans find common ground in democratic deliberation. Science 386, eadq2852 (2024).

    Article 
    CAS 
    PubMed 

    Google Scholar
     

  • Goldstein, J. A. et al. How persuasive is AI-generated propaganda? PNAS Nexus 3, pgae034 (2024).

    Article 
    PubMed 
    PubMed Central 

    Google Scholar
     

  • Fisher, J. et al. Biased LLMs can influence political decision-making. In Proc. 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers) (eds Che, W. et al.) 6559–6607 (Association for Computational Linguistics, 2025).

  • Saenger, T.R. et al. AutoPersuade: a framework for evaluating and explaining persuasive arguments. In Proc. 2024 Conference on Empirical Methods in Natural Language Processing (eds Al-Onaizan, Y., Bansal, M. & Chen, Y.-N.) 16325–16342 (Association for Computational Linguistics, 2024).

  • Islas-Carmona, J. O., Gutiérrez-Cortés, F. I. & Arribas-Urrutia, A. Disinformation and political propaganda: an exploration of the risks of artificial intelligence. Explor. Media Ecol. 23, 105–120 (2024).


    Google Scholar
     

  • Woolley, S. Manufacturing Consensus: Understanding Propaganda in the Era of Automation and Anonymity (Yale Univ. Press, 2023).

  • Broockman, D. & Kalla, J. Durably reducing transphobia: a field experiment on door-to-door canvassing. Science 352, 220–224 (2016).

    Article 
    ADS 
    CAS 
    PubMed 

    Google Scholar
     

  • Roghanizad, M. M. & Bohns, V. K. Ask in person: you’re less persuasive than you think over email. J. Exp. Soc. Psychol. 69, 223–226 (2017).

    Article 

    Google Scholar
     

  • Buyl, M. et al. Large language models reflect the ideology of their creators. npj Artif. Intell. 2, 7 (2026).

    Article 
    PubMed 
    PubMed Central 

    Google Scholar
     

  • Guey, W. et al. Mapping geopolitical bias in 11 large language models: a bilingual, dual-framing analysis of US-China tensions. Preprint at https://arxiv.org/abs/2503.23688 (2025).

  • McCarthy, S. DeepSeek is giving the world a window into Chinese censorship and information control. CNN https://edition.cnn.com/2025/01/29/china/deepseek-ai-china-censorship-moderation-intl-hnk (29 January 2025).

  • Ouyang, Y., Nellis, S. and Tong, Q. DeepSeek hit by cyberattack as users flock to Chinese AI startup. Reuters https://www.reuters.com/technology/artificial-intelligence/chinese-ai-startup-deepseek-overtakes-chatgpt-apple-app-store-2025-01-27/ (27 January 2025).

  • Kachwala, Z. Musk’s xAI updates Grok chatbot after ‘white genocide’ comments. Reuters https://www.reuters.com/business/musks-xai-updates-grok-chatbot-after-white-genocide-comments-2025-05-17/ (17 May 2025).

  • O’Brien, M. Google says its AI image-generator would sometimes ‘overcompensate’ for diversity. Associated Press https://apnews.com/article/google-gemini-ai-chatbot-imagegenerator-race-c7e14de837aa65dd84f6e7ed6cfc4f4b (23 February 2024).

  • O’Brien, M. Elon Musk’s AI company says Grok chatbot focus on South Africa’s racial politics was ‘unauthorized’. Associated Press https://apnews.com/article/grok-ai-south-africa-64ce5f240061ca0b88d5af4c424e1f3b (16 May 2025).

  • Price, M. E. Media and Sovereignty: The Global Information Revolution and its Challenge to State Power (MIT Press, 2002).

  • Hallin, D. C. & Mancini, P. Comparing Media Systems: Three Models of Media and Politics (Cambridge Univ. Press, 2004).

  • Gururangan, S. et al. Don’t stop pretraining: adapt language models to domains and tasks. In Proc. 58th Annual Meeting of the Association for Computational Linguistics (eds Jurafsky, D. et al.) 8342–8360 (Association for Computational Linguistics, 2020).

  • Bender, E. M. et al. On the dangers of stochastic parrots: can language models be too big? In Proc. 2021 ACM Conference on Fairness, Accountability, and Transparency 610–623 (Association for Computing Machinery, 2021).

  • Kreutzer, J. et al. Quality at a glance: an audit of web-crawled multilingual datasets. Trans. Assoc. Comput. Linguist. 10, 50–72 (2022).

    Article 

    Google Scholar
     

  • Blodgett, S. L. et al. Language (technology) is power: a critical survey of “bias” in NLP. In Proc. Annual Meeting of the Association for Computational Linguistics (eds Jurafsky, D. et al.) 5454–5476 (Association for Computational Linguistics, 2020).

  • Ouyang, L. et al. Training language models to follow instructions with human feedback. Adv. Neural Inf. Process. Syst. 35, 27730–27744 (2022).


    Google Scholar
     

  • Bai, Y. et al. Constitutional AI: harmlessness from AI feedback. Preprint at https://arxiv.org/abs/2212.08073 (2022).

  • Bulté, B. & Terryn, A. R. LLMs and cultural values: the impact of prompt language and explicit cultural framing. Comput. Linguist. https://doi.org/10.1162/COLI.a.583 (2026).

  • Lu, J. G., Song, L. L. & Zhang, L. D. Cultural tendencies in generative AI. Nat. Hum. Behav. 9, 2360–2369 (2025).

  • Kay, M., Matuszek, C. & Munson, S. A. Unequal representation and gender stereotypes in image search results for occupations. In Proc. 33rd Annual ACM Conference on Human Factors in Computing Systems (eds Begole, B. et al.) 3819–3828 (Association for Computing Machinery, 2015).

  • Noble, S. U. Algorithms of Oppression: How Search Engines Reinforce Racism (New York Univ. Press, 2018).

  • Broussard, M. More than a Glitch: Confronting Race, Gender, and Ability Bias in Tech (MIT Press, 2023).

  • Benjamin, R. Race after Technology: Abolitionist Tools for the New Jim Code (John Wiley & Sons, 2019).

  • Buolamwini, J. & Gebru, T. Gender shades: intersectional accuracy disparities in commercial gender classification. In Conference on Fairness, Accountability and Transparency (eds Friedler, S. A. & Wilson, C.) 77–91 (PMLR, 2018).

  • Barocas, S. & Selbst, A. D. Big data’s disparate impact. Calif. L. Rev. 104, 671 (2016).


    Google Scholar
     

  • Sheng, E. et al. The woman worked as a babysitter: on biases in language generation. In Proc. 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP) (eds Inui, K. et al.) 3407–3412 (Association for Computational Linguistics, 2019).

  • Field, A. et al. A survey of race, racism, and anti-racism in NLP. In Proc. 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers) (eds Zong, C. et al.) 1905–1925 (Association for Computational Linguistics, 2021).

  • Metaxa, D. et al. An image of society: gender and racial representation and impact in image search results for occupations. Proc. ACM Hum. Comp. Interact. 5, 1–23 (2021).

    Article 

    Google Scholar
     

  • Kotek, H., Dockum, R. & Sun, D. Gender bias and stereotypes in large language models. In Proc. ACM Collective Intelligence Conference (eds Bernstein, M. S. et al.) 12–24 (Association for Computing Machinery, 2023).

  • Omiye, J. A. et al. Large language models propagate race-based medicine. NPJ Digit. Med. 6, 195 (2023).

    Article 
    PubMed 
    PubMed Central 

    Google Scholar
     

  • Jowett, G. S. & O’Donnell, V. Propaganda & Persuasion (Sage, 2018).

  • Peisakhin, L. & Rozenas, A. Electoral effects of biased media: Russian television in Ukraine. Am. J. Polit. Sci. 62, 535–550 (2018).

    Article 

    Google Scholar
     

  • Selb, P. & Munzert, S. Examining a most likely case for strong campaign effects: Hitler’s speeches and the rise of the Nazi party, 1927-1933. Am. Polit. Sci. Rev. 112, 1050–1066 (2018).

    Article 

    Google Scholar
     

  • Rozenas, A. & Stukal, D. How autocrats manipulate economic news: evidence from Russia’s state-controlled television. J. Polit. 81, 982–996 (2019).

    Article 

    Google Scholar
     

  • Huang, H. Propaganda as signaling. Comp. Polit. 47, 419–444 (2015).

    Article 

    Google Scholar
     

  • Voigtländer, N. & Voth, H.-J. Nazi indoctrination and anti-Semitic beliefs in Germany. Proc. Natl Acad. Sci. USA 112, 7931–7936 (2015).

    Article 
    ADS 
    PubMed 
    PubMed Central 

    Google Scholar
     

  • King, G., Pan, J. & Roberts, M. E. How the Chinese government fabricates social media posts for strategic distraction, not engaged argument. Am. Polit. Sci. Rev. 111, 484–501 (2017).

    Article 

    Google Scholar
     

  • Stukal, D. et al. Why botter: how pro-government bots fight opposition in Russia. Am. Polit. Sci. Rev. 116, 843–857 (2022).

    Article 

    Google Scholar
     

  • Farzam, A. et al. Opinion manipulation on Farsi Twitter. Sci. Rep. 13, 333 (2023).

    Article 
    ADS 
    CAS 
    PubMed 
    PubMed Central 

    Google Scholar
     

  • Waight, H. et al. The decade-long growth of government-authored news media in China under Xi Jinping. Proc. Natl Acad. Sci. USA 122, e2408260122 (2025).

    Article 
    CAS 
    PubMed 
    PubMed Central 

    Google Scholar
     

  • Yang, E. & Roberts, M. E. Censorship of online encyclopedias: implications for NLP models. In Proc. 2021 ACM Conference on Fairness, Accountability, and Transparency 537–548 (Association for Computing Machinery, 2021).

  • Zhou, D. & Zhang, Y. Political biases and inconsistencies in bilingual GPT models — the cases of the US and China. Sci. Rep. 14, 25048 (2024).

    Article 
    ADS 
    CAS 
    PubMed 
    PubMed Central 

    Google Scholar
     

  • Ahmed, M. & Knockel, J. Extended abstract: the impact of online censorship on LLMs. Free and Open Communications on the Internet https://www.petsymposium.org/foci/2024/foci-2024-0006.pdf (2024).

  • Urman, A. & Makhortykh, M. The silence of the LLMs: cross-lingual analysis of guardrail-related political bias and false information prevalence in ChatGPT, Google Bard (Gemini), and Bing Chat. Telemat. Informat. 96, 102211 (2025).

    Article 

    Google Scholar
     

  • Spirling, A. & Stewart, B. M. What good is a regression? Inference to the best explanation and the practice of political science research. J. Polit. 87, 1587–1599 (2025).

    Article 

    Google Scholar
     

  • Reporters Without Borders. World Press Freedom Index https://rsf.org/en/index (2024).

  • Shambaugh, D. in Critical Readings on the Communist Party of China (4 Vols. Set) (ed. Brødsgaard, K. E.) 713–751 (Brill, 2017).

  • Brady, A.M. Marketing Dictatorship: Propaganda and Thought Work in Contemporary China (Rowman & Littlefield, 2009).

  • Stockmann, D. Media Commercialization and Authoritarian Rule in China (Cambridge Univ. Press, 2013).

  • Liang, F., Chen, Y. & Zhao, F. The platformization of propaganda: how Xuexi Qiangguo expands persuasion and assesses citizens in China. Int. J. Commun. 15, 20 (2021).


    Google Scholar
     

  • Lu, Y. & Pan, J. Capturing clicks: how the Chinese government uses clickbait to compete for visibility. Polit. Commun. 38, 23–54 (2021).

    Article 

    Google Scholar
     

  • Repnikova, M. & Fang, K. Digital media experiments in China: ‘revolutionizing’ persuasion under Xi Jinping. China Q. 239, 679–701 (2019).

    Article 

    Google Scholar
     

  • Esarey, A. Winning hearts and minds? Cadres as microbloggers in China. J. Curr. Chinese Aff. 44, 69–103 (2015).

    Article 

    Google Scholar
     

  • Qin, B., Strömberg, D. & Wu, Y. Media bias in China. Am. Econom. Rev. 108, 2442–2476 (2018).

    Article 

    Google Scholar
     

  • Pan, J., Shao, Z. & Xu, Y. How government-controlled media shifts policy attitudes through framing. Polit. Sci. Res. Methods 10, 317–332 (2022).

    Article 

    Google Scholar
     

  • Zhang, Z. et al. Unveiling linguistic regions in large language models. In Proc. 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers) (eds Ku, L.-W., Martins, A. & Srikumar, V.) 6228–6247 (Association for Computational Linguistics, 2024).

  • Qi, J., Fernández, R. & Bisazza, A. Cross-lingual consistency of factual knowledge in multilingual language models. In Proc. 2023 Conference on Empirical Methods in Natural Language Processing (eds Bouamor, H., Pino, J. & Bali, K.) 10650–10666 (Association for Computational Linguistics, 2023).

  • Li, B., Haider, S. & Callison-Burch, C. This land is your, my land: evaluating geopolitical bias in language models through territorial disputes. In Proc. 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers) (eds Duh, K., Gomez, H. & Bethard, S.) 3855–3871 (Association for Computational Linguistics, 2024).

  • Wendler, C. et al. Do llamas work in English? On the latent language of multilingual transformers. In Proc. 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers) (eds Ku, L.-W., Martins, A. & Srikumar, V.) 15366–15394 (Association for Computational Linguistics, 2024).

  • Durmus, E. et al. Towards measuring the representation of subjective global opinions in language models. In 1st Conference on Language Modeling https://openreview.net/pdf?id=zl16jLb91v (COLM, 2024).

  • Shayegani, E. et al. Survey of vulnerabilities in large language models revealed by adversarial attacks. Preprint at https://arxiv.org/abs/2310.10844 (2023).

  • Roberts, M. Censored: Distraction and Diversion Inside China’s Great Firewall (Princeton Univ. Press, 2018).

  • Ishihara, S. & Takahashi, H. Quantifying memorization and detecting training data of pre-trained language models using Japanese newspaper. In Proc. 17th International Natural Language Generation Conference (eds Mahamood, S., Le Minh, N. & Ippolito, D.) 165–179 (Association for Computational Linguistics, 2024).

  • Fulay, S. et al. On the relationship between truth and political bias in language models. In Proc. 2024 Conference on Empirical Methods in Natural Language Processing (eds Al-Onaizan, Y., Bansal, M. & Chen, Y.-N.) 9004–9018 (Association for Computational Linguistics, 2024).

  • Nguyen, T. et al. CulturaX: a cleaned, enormous, and multilingual dataset for large language models in 167 languages. In Proc. 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024) (eds Calzolari, N. et al.) 4226–4237 (ELRA and ICCL, 2024).

  • Truex, R. Focal points, dissident calendars, and preemptive repression. J. Confl. Resolut. 63, 1032–1052 (2019).

    Article 

    Google Scholar
     

  • Carter, E. B. & Carter, B. L. When autocrats threaten citizens with violence: evidence from China. Br. J. Polit. Sci. 52, 671–696 (2022).

  • Schlessinger, J. et al. Exposing the obscured influence of state-controlled media via causal inference of quotation propagation. Sci. Rep. 15, 1110 (2025).

    Article 
    ADS 
    CAS 
    PubMed 
    PubMed Central 

    Google Scholar
     

  • Zhao, W. et al. WildChat: 1M ChatGPT interaction logs in the wild. In 12th International Conference on Learning Representations https://openreview.net/forum?id=Bl8u7ZRlbM (ICLR, 2024).

  • Trussler, M. & Soroka, S. Consumer demand for cynical and negative news frames. Int. J. Press Polit. 19, 360–379 (2014).

    Article 

    Google Scholar
     

  • Arango-Kure, M., Garz, M. & Rott, A. Bad news sells: the demand for news magazines and the tone of their covers. J. Media Econom. 27, 199–214 (2014).

    Article 

    Google Scholar
     

  • Christin, A. Counting clicks: quantification and variation in web journalism in the United States and France. Am. J. Sociol. 123, 1382–1415 (2018).

    Article 

    Google Scholar
     

  • O’Neil, C. Weapons of Math Destruction: How Big Data Increases Inequality and Threatens Democracy (Crown, 2017).

  • Fourcade, M. & Healy, K. The Ordinal Society (Harvard Univ. Press, 2024).

  • Gillespie, T. Custodians of the Internet: Platforms, Content Moderation, and the Hidden Decisions that Shape Social Media (Yale Univ. Press, 2018).

  • Yang, E. & Roberts, M. E. The authoritarian data problem. J. Democr. 34, 141–150 (2023).

    Article 

    Google Scholar
     

  • Wang, H. & Sparks, C. Chinese newspaper groups in the digital era: the resurgence of the party press. J. Commun. 69, 94–119 (2019).

    Article 
    ADS 

    Google Scholar
     

  • Raffel, C. et al. Exploring the limits of transfer learning with a unified text-to-text transformer. J. Mach. Learn. Res. 21, 1–67 (2020).


    Google Scholar
     

  • Scheible, R. et al. GottBERT: a pure German language model. In Proc. 2024 Conference on Empirical Methods in Natural Language Processing (eds Al-Onaizan, Y., Bansal, M. & Chen, Y.-N.) 21237–21250 (Association for Computational Linguistics, 2024).

  • Shalumov, V. & Haskey, H. Hero: ROBERTa and longformer Hebrew language models. Preprint at https://arxiv.org/abs/2304.11077 (2023).

  • Serrano, A. V. et al. RigoBERTa: a state-of-the-art language model for Spanish. Preprint at https://arxiv.org/abs/2205.10233 (2022).

  • Shliazhko, O. et al. mGPT: few-shot learners go multilingual. Trans. Assoc. Comput. Linguist. 12, 58–79 (2024).

    Article 

    Google Scholar
     

  • Mandal, P. K. & Mahto, R. An FNet based auto encoder for long sequence news story generation. Preprint at https://arxiv.org/abs/2211.08295 (2022).

  • Boumans, J. et al. The agency makes the (online) news world go round: the impact of news agency content on print and online news. Int. J. Commun. 12, 22 (2018).


    Google Scholar
     

  • Cagé, J., Hervé, N. & Viaud, M.-L. The production of information in an online world. Rev. Econom. Stud. 87, 2126–2164 (2020).

    Article 

    Google Scholar
     

  • Nicholls, T. Detecting textual reuse in news stories, at scale. Int. J. Commun. 13, 4173–4197 (2019).

  • Gao, L. et al. The pile: an 800gb dataset of diverse text for language modeling. Preprint at https://arxiv.org/abs/2101.00027 (2020).

  • Carlini, N. et al. Quantifying memorization across neural language models. In 11th International Conference on Learning Representations https://openreview.net/forum?id=TatRHT_1cK (ICLR, 2023).

  • Touvron, H. et al. Llama 2: open foundation and fine-tuned chat models. Preprint at https://arxiv.org/abs/2307.09288 (2023).

  • Chen, L. et al. AlpaGasus: training a better alpaca with fewer data. In International Conference on Learning Representations (eds Kim, B. et al.) https://proceedings.iclr.cc/paper_files/paper/2024/hash/9543942c237ded1b39b1fd37259ff88e-Abstract-Conference.html (ICLR, 2024).

  • Hu, E. J. et al. LoRA: low-rank adaptation of large language models. In 10th International Conference on Learning Representations https://openreview.net/forum?id=nZeVKeeFYf9 (ICLR, 2022).

  • Kirkpatrick, J. et al. Overcoming catastrophic forgetting in neural networks. Proc. Natl Acad. Sci. USA 114, 3521–3526 (2017).

    Article 
    ADS 
    CAS 
    PubMed 
    PubMed Central 

    Google Scholar
     

  • Leybzon, D. D. & Kervadec, C. Learning, forgetting, remembering: insights from tracking LLM memorization during training. In Proc. 7th BlackboxNLP Workshop: Analyzing and Interpreting Neural Networks for NLP (eds Belinkov, Y. et al.) 43–57 (Association for Computational Linguistics, 2024).

  • Mahomed, Y. et al. Auditing GPT’s content moderation guardrails: can ChatGPT write your favorite TV show? In 2024 ACM Conference on Fairness, Accountability, and Transparency 660–686 (Association for Computing Machinery, 2024).

  • Xu, B. NLP Chinese Corpus: large scale Chinese Corpus for NLP (version 1.0). Zenodo https://doi.org/10.5281/zenodo.3402023 (2019).

  • Egami, N. et al. Using imperfect surrogates for downstream inference: design-based supervised learning for social science applications of large language models. Adv. Neural Inf. Process. Syst. 36, 68589–68601 (2024).


    Google Scholar
     

  • Eberhard, D. M., Simons, G. F. & Fennig, C. D. (eds). Ethnologue: Languages of the World (SIL International, 2024).

  • Source link

    Leave a Reply

    Your email address will not be published. Required fields are marked *