Research Article | | Peer-Reviewed

A Sociopolitical Approach to Disinformation and AI: Concerns, Responses and Challenges

Received: 2 October 2024     Accepted: 17 October 2024     Published: 31 October 2024
Views:       Downloads:
Abstract

International organizations classify disinformation as one of the main threats to democracy and institutions for more than a decade. Digital technologies reinvent and profoundly transform modern lifestyles, citizens’ and business environments. AI is bringing a new disruption in the way we access knowledge and create, spread and understand information. It can also blur the lines between real information and manipulated information with the emergence of ‘Fake News’, automatic networks’ cross referencing, and ‘Deep Fakes’. AI systems enhance the potential for creating realistic fake content and targeted disinformation campaigns. Disinformation goes beyond simple rumors to deliberately deceive and distort evidence-based information through fabricated data. European institutions have also recently focused on the identification of disinformation linked to FIMI: Foreign Information Manipulation and Interference. The article identifies trends and concerns related to disinformation and AI. It explores the perception of disinformation, its impacts, and responses including the EU AI Act and online Platforms’ policies. It provides a first analytical approach to the topic based on the current debates by researchers, the first findings of our 2024 surveys, interviews and the analysis of hundreds of online fake news items. It attempts to understand how citizens and stakeholders perceive disinformation and identifies possible impacts. It also analyzes the current challenges and constraints, opportunities and limitations to tackle manipulation and interference. The article considers the current processes, and impacts of disinformation (2), the presentation of the main findings of our online survey on the perceptions of disinformation (3), the current EU regulatory responses (4) and the Discussion Points (5). We argue in this article that there is a gigantic change in the way that we access information, but that the responses to disinformation are still at an early stage. The article also demonstrates that there is an increased awareness in European countries about the impacts of disinformation, but also a gap between the ability to identify "fake news" and disinformation, and a limited understanding of the processes, threats, and actors involved in spreading disinformation.

Published in Journal of Political Science and International Relations (Volume 7, Issue 4)
DOI 10.11648/j.jpsir.20240704.11
Page(s) 75-88
Creative Commons

This is an Open Access article, distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution and reproduction in any medium or format, provided the original work is properly cited.

Copyright

Copyright © The Author(s), 2024. Published by Science Publishing Group

Keywords

Disinformation, Fake News, Misinformation, Artificial Intelligence, Manipulation, Interference, Social Media

References
[1] AI4DEBUNK 2024, ‘Towards of Theory Framework’, AI4debunk, Riga, 13 March 2024.
[2] Antoniuk, D. (2023, November 8). Russian ‘influence-for-hire’ firms spread propaganda in Latin America: US State Department. The Record by Recorded Future.
[3] Art, S.: Media literacy and critical thinking. International Journal of Media and Information Literacy, 3(2), 2018. 66-71.
[4] Bauer M., Cahlíková J., Chytilová J., Roland G., Želinský T.: Shifting Punishment onto Minorities: Experimental Evidence of Scapegoating, The Economic Journal, Volume 133, Issue 652, May 2023, 1626–1640,
[5] Bergmanis-Koräts G., Arhippainen M. et al. Virtual Manipulation Brief. Highjacking Reality. The increased role of Generative AI in Russian Propaganda. NATO Stratcom. 2024
[6] Betz H. G., Oswald M. L. Emotional Mobilization: The affective Underpinnings of Right -Wing Populist Party Support., Palgrave Handbook of Populism 2021. 115-143.
[7] Bollmann, H. S., & Gibeon, G. (2022). The spread of hacked materials on Twitter: A threat to democracy? A case study of the 2017 Macron Leaks (Doctoral dissertation, Hertie School).
[8] Bontridder N. and Poullet Y. The role of artificial intelligence in disinformation. Data & Policy, 3: 2021, e32.
[9] Butcher, P., & Neidhardt, A. H.: Fear and lying in the EU: Fighting disinformation on migration with alternative narratives. Foundation for European Progressive Studies, 2020.
[10] Casten Stahl: On the Difference or Equality of Information, Misinformation, and Disinformation: A Critical Research Perspective. Informing Science: The International Journal of an Emerging Transdiscipline Volume 9, 2006, 083-096
[11] Charillon F.. Guerres d’influence. Odile Jacob 2018.
[12] Cull. N. J., 2009, ‘Public Diplomacy: lessons from the past’ USC center of public diplomacy.
[13] Darwin Rusdin, D., Mukminatien, N., Suryati, N., Laksmi, E. D., & Marzuki: Critical thinking in the AI era: An exploration of EFL students’ perceptions, benefits, and limitations. Cogent Education, 11(1), 2290342. 2024.
[14] Dauksas V., Venclauskiené L., Urbanaviciuté K., Friedman O: War on all fronts: How the Kremlin’s Media Ecosystem broadcasts the war in Ukraine. NATO Stratcom
[15] Deutsch M.: Trust and Suspicion, Conflict Resolution Number 2 (Vol8) 1958.
[16] ENISA European Union Agency for Cybersecurity (i), Lella, I., Ciobanu, C., Tsekmezoglou, E. (2023). ENISA threat landscape 2023: July 2022 to June 2023, (I. Lella, editor, C. Ciobanu, editor, M. Theocharidou, editor, E. Magonara, editor, A. Malatras, editor, R. Svetozarov Naydenov, editor, E. Tsekmezoglou, edito). Retrieved from:
[17] European Union Agency for Cybersecurity (ii), Tsekmezoglou, E., Lella, I., Malatras, A. et al., ENISA threat landscape for DoS attack – January 2022 to August 2023, European Union Agency for Cybersecurity, 2023, retrieved from:
[18] European Commission, 2018a, A Multi-dimensional Approach to Disinformation: Report of the Independent High-Level Group on Fake News and Online Disinformation. Directorate-General for Communication Networks, Content and Technology. Available at
[19] European Commission, 2018b, Code of Practice on Disinformation. Available at
[20] European Commission, 2018c, Synopsis Report of the Public Consultation on Fake News and Online Disinformation.
[21] European Commission (2018d) Tackling Online Disinformation: A European Approach (Communication) COM (2018) 236 final.
[22] European Commission (2020a) Assessment of the Code of Practice on Disinformation —Achievements and areas for further improvement. Commission Staff working document (SWD (2020) 180 final).
[23] European Commission (2020b) European Democracy Action Plan (Communication) COM (2020) 790 final.
[24] European Commission (2021) Guidance on Strengthening the Code of Practice on Disinformation (COM (2021) 262 final).
[25] Regulation (EU) 2024/1689 of the European Parliament and of the Council of 13 June 2024 laying down harmonised rules on artificial intelligence and amending Regulations (EC) No 300/2008, (EU) No 167/2013, (EU) No 168/2013, (EU) 2018/858, (EU) 2018/1139 and (EU) 2019/2144 and Directives 2014/90/EU, (EU) 2016/797 and (EU) 2020/1828 (Artificial Intelligence Act) (Text with EEA relevance)
[26] Regulation (EU) 2022/2065 of the European Parliament and of the Council of 19 October 2022 on a Single Market For Digital Services and amending Directive 2000/31/EC (Digital Services Act)
[27] Fine G. A.: Rumor, Trust and Civil Society: Collective Memory and Cultures of Judgment. Diogenes 2007, 54 (1): 5-18.
[28] Foreign Affairs Review, 2017 ‘The meaning of sharp power: How authoritarian States project Influence’, Foreign Affairs Review, 16th November 2017.
[29] Gaborit P.: Restaurer la confiance après un conflit civil, L’Harmattan 2009 a.
[30] Gaborit P.: La confiance après un conflit ou la confiance désenchantée, in Bertho A., Gaumont-Prat H. et Serry H. Colloque international La confiance et le conflit, Université Paris Vincennes Saint Denis 2009 b.
[31] Girard R. The Scapegoat, Johns Hopkins University Press, 1986.
[32] Goodhart D.: The future to somewhere: The populist revolt and the future of politics. London, Hurst and Company, 2017. 9781849047999.
[33] G. Rodriguez-Pose. A.: The revenges of the places that don’t matter- and what to do about it. Cambridge Journal of Regions, Economy and Society, II (I), 2017: 189-201.
[34] Grabner-Kräuter S.: Empiral Research in Online Trust. A Review and Critical Assessment. International Journal of Human-Computer Study. 2003
[35] Grinberg, N., Joseph, K., Friedland, L., Swire-Thompson, B., & Lazer, D.: Fake news on Twitter during the 2016 US presidential election. Science, 363(6425), 2019. 374-378.
[36] Kertysova K.: Artificial Intelligence and Disinformation How AI Changes the Way Disinformation is Produced, Disseminated, and Can Be Countered, security and human rights 29. 2018. 55-81.
[37] Lloyd J. and Toogood L. (published with I. B. Tauris): Journalism and PR: News Media and Public Relations in the Digital Age. Oxford University and Reuteurs institute. 2015.
[38] Hamm, J. A., van der Werff, L., Osuna, A. I., Blomqvist, K., Blount-Hill, K. L., Gillespie, N., … Tomlinson, E. C.: Capturing the conversation of trust research. Journal of Trust Research, 14(1), 1–7. 2024
[39] Hardin R. (Ed): Trust and Trusworthiness. New York, Russel Sage foundation editions, collection on trust, volume 4, 2002.
[40] Hardin R. (Ed): Distrust, NYC, Russell Sage Foundation. 2004.
[41] Haiduchyk T., Shevtsov A., Bergmanis-Koräts G. AI in Precision Persuasion: Unveiling Tactics and Risks on Social Media. NATO Stratcom 2024
[42] Hersh M. A.: Barriers to ethical behaviour and stability: Stereotyping and scapegoating as pretexts for avoiding responsibility, Annual Reviews in Control, Volume 37, Issue 2, 2013, 365-381,
[43] King K., Wang b. Diffusion of real versus misinformation during a crisis event: A big data driven approach. International Journal of Information Management. 71. 2023
[44] Kueng L.: Hearts and Minds: Harnessing Leadership, Culture, and Talent to Really Go Digital, Oxford University, Reuteurs Institute, 2020.
[45] Kunelius R., Heikkilä H., Russell A. and Yagodin D. (eds) (published with I. B. Tauris):, Journalism and the NSA Revelations: Privacy, Security and the Press, 2017.
[46] Luhmann, N: Trust and Power: Two Works by Niklas Luhmann. Translation of German originals Vertrauen 1968 and Macht 1975. Chichester: John Wiley. 1979.
[47] Małecka, A. (2024). Non-State Actors in Nation-State Cyber Operations. Rocznik Bezpieczeństwa Międzynarodowego, 18(1), 45-64.
[48] Mont’Alverne C., Badrinathan S., Ross Arguedas A., Toff B., Fletcher R., and Kleis Nielsen R.: The Trust Gap: How and Why News on Digital Platforms Is Viewed More Sceptically Versus News in General, Reuters Institute, 2022
[49] Moravcsik A.. Taking preferences seriously: A Liberal Theory of International politics’, International Organization, vol 4, n°51, fall 1997, p 513-533.
[50] Newman N.: Digital News Project: Journalism, Media and Technology: Trends and Prediction. Oxford University, Reuters Institute, 2024.
[51] Persily N. and Tucker J. A: Social Media and Democracy The State of the Field, Prospects for Reform, Cambridge University Press, 2021.
[52] Putnam R.: Making Democracy work: Civic traditions in Modern Italy, Princeton University Press 1993.
[53] Romero Vincente A et al.. ‘Coordinated Inauthentic Behavior’ EU Disinfo Lab 2024.
[54] Sessa M. G., 2023, EU Disinfolab ‘Connecting the Disinformation Dots’ Friedrich Nauman Foundation.
[55] Sessa M. G., Miguel R. The Doppelganger Case: Assessment of Platform Regulation on the EU Disinformation Environment. NATO Stratcom. 2024.
[56] Seligman A.: The problem of Trust, Princeton, Princeton University Press. 1997. 9780691050201.
[57] Shahbazi M., Bunker D. Social media Trust: Fighting misinformation in the time of crisis. Information Journal of Information Management. 77. 2024.
[58] Six F. E.; Latusek D.: Distrust: A critical review exploring a universal distrust sequence, Journal of Trust Research, 13: 1, 1-23, 2024
[59] Scheirer W. A Review of A History of Fake Things on the Internet. Stanford University Press 2023
[60] Smith, R. B., Perry, M., & Smith, N. N.: Fake News’ in ASEAN: Legislative responses. Journal of ASEAN Studies, 9(2), 2021. 117-137.
[61] Smuha, Nathalie A.. "Beyond the individual: governing AI’s societal harm". Internet Policy Review 10. 3 2021
[62] Stahl B. C., On the Difference or Equality of Information, Misinformation, and Disinformation: A Critical Research Perspective’ Informing social science, Vol 9, 2006.
[63] Sztompka P.: Trust a sociological theory, New York, Cambridge University Press. 2000.
[64] Tilly C.: Trust and Rule, Cambridge University Press. 2005.
[65] Wade M.. Psychographics: The Behavioural Analysis That Helped Cambridge Analytica Know Voters’ Minds. The Conversation, March 21, 2018,
[66] Whyte C. Deepfake news: AI-enabled disinformation as a multi-level public policy challenge, Journal of Cyber Policy, 5: 2, 2020; 199-217,
[67] Witzel L: 5 Things You Must Know Now About the Coming EU AI Regulation,
Cite This Article
  • APA Style

    Gaborit, P. (2024). A Sociopolitical Approach to Disinformation and AI: Concerns, Responses and Challenges. Journal of Political Science and International Relations, 7(4), 75-88. https://doi.org/10.11648/j.jpsir.20240704.11

    Copy | Download

    ACS Style

    Gaborit, P. A Sociopolitical Approach to Disinformation and AI: Concerns, Responses and Challenges. J. Polit. Sci. Int. Relat. 2024, 7(4), 75-88. doi: 10.11648/j.jpsir.20240704.11

    Copy | Download

    AMA Style

    Gaborit P. A Sociopolitical Approach to Disinformation and AI: Concerns, Responses and Challenges. J Polit Sci Int Relat. 2024;7(4):75-88. doi: 10.11648/j.jpsir.20240704.11

    Copy | Download

  • @article{10.11648/j.jpsir.20240704.11,
      author = {Pascaline Gaborit},
      title = {A Sociopolitical Approach to Disinformation and AI: Concerns, Responses and Challenges
    },
      journal = {Journal of Political Science and International Relations},
      volume = {7},
      number = {4},
      pages = {75-88},
      doi = {10.11648/j.jpsir.20240704.11},
      url = {https://doi.org/10.11648/j.jpsir.20240704.11},
      eprint = {https://article.sciencepublishinggroup.com/pdf/10.11648.j.jpsir.20240704.11},
      abstract = {International organizations classify disinformation as one of the main threats to democracy and institutions for more than a decade. Digital technologies reinvent and profoundly transform modern lifestyles, citizens’ and business environments. AI is bringing a new disruption in the way we access knowledge and create, spread and understand information. It can also blur the lines between real information and manipulated information with the emergence of ‘Fake News’, automatic networks’ cross referencing, and ‘Deep Fakes’. AI systems enhance the potential for creating realistic fake content and targeted disinformation campaigns. Disinformation goes beyond simple rumors to deliberately deceive and distort evidence-based information through fabricated data. European institutions have also recently focused on the identification of disinformation linked to FIMI: Foreign Information Manipulation and Interference. The article identifies trends and concerns related to disinformation and AI. It explores the perception of disinformation, its impacts, and responses including the EU AI Act and online Platforms’ policies. It provides a first analytical approach to the topic based on the current debates by researchers, the first findings of our 2024 surveys, interviews and the analysis of hundreds of online fake news items. It attempts to understand how citizens and stakeholders perceive disinformation and identifies possible impacts. It also analyzes the current challenges and constraints, opportunities and limitations to tackle manipulation and interference. The article considers the current processes, and impacts of disinformation (2), the presentation of the main findings of our online survey on the perceptions of disinformation (3), the current EU regulatory responses (4) and the Discussion Points (5). We argue in this article that there is a gigantic change in the way that we access information, but that the responses to disinformation are still at an early stage. The article also demonstrates that there is an increased awareness in European countries about the impacts of disinformation, but also a gap between the ability to identify "fake news" and disinformation, and a limited understanding of the processes, threats, and actors involved in spreading disinformation.
    },
     year = {2024}
    }
    

    Copy | Download

  • TY  - JOUR
    T1  - A Sociopolitical Approach to Disinformation and AI: Concerns, Responses and Challenges
    
    AU  - Pascaline Gaborit
    Y1  - 2024/10/31
    PY  - 2024
    N1  - https://doi.org/10.11648/j.jpsir.20240704.11
    DO  - 10.11648/j.jpsir.20240704.11
    T2  - Journal of Political Science and International Relations
    JF  - Journal of Political Science and International Relations
    JO  - Journal of Political Science and International Relations
    SP  - 75
    EP  - 88
    PB  - Science Publishing Group
    SN  - 2640-2785
    UR  - https://doi.org/10.11648/j.jpsir.20240704.11
    AB  - International organizations classify disinformation as one of the main threats to democracy and institutions for more than a decade. Digital technologies reinvent and profoundly transform modern lifestyles, citizens’ and business environments. AI is bringing a new disruption in the way we access knowledge and create, spread and understand information. It can also blur the lines between real information and manipulated information with the emergence of ‘Fake News’, automatic networks’ cross referencing, and ‘Deep Fakes’. AI systems enhance the potential for creating realistic fake content and targeted disinformation campaigns. Disinformation goes beyond simple rumors to deliberately deceive and distort evidence-based information through fabricated data. European institutions have also recently focused on the identification of disinformation linked to FIMI: Foreign Information Manipulation and Interference. The article identifies trends and concerns related to disinformation and AI. It explores the perception of disinformation, its impacts, and responses including the EU AI Act and online Platforms’ policies. It provides a first analytical approach to the topic based on the current debates by researchers, the first findings of our 2024 surveys, interviews and the analysis of hundreds of online fake news items. It attempts to understand how citizens and stakeholders perceive disinformation and identifies possible impacts. It also analyzes the current challenges and constraints, opportunities and limitations to tackle manipulation and interference. The article considers the current processes, and impacts of disinformation (2), the presentation of the main findings of our online survey on the perceptions of disinformation (3), the current EU regulatory responses (4) and the Discussion Points (5). We argue in this article that there is a gigantic change in the way that we access information, but that the responses to disinformation are still at an early stage. The article also demonstrates that there is an increased awareness in European countries about the impacts of disinformation, but also a gap between the ability to identify "fake news" and disinformation, and a limited understanding of the processes, threats, and actors involved in spreading disinformation.
    
    VL  - 7
    IS  - 4
    ER  - 

    Copy | Download

Author Information
  • Sections