Skip to main navigation Skip to search Skip to main content

Artificial Intelligence for Detecting Electoral Disinformation on Social Media: Models, Datasets, and Evaluation

    Research output: Contribution to journalReview articlepeer-review

    Abstract

    During elections, information manipulation on social media has accelerated the use of artificial intelligence, yet the evidence is difficult to interpret without an integrated view of methods, data, and evaluation. We mapped 557 English-language journal articles from Scopus and Web of Science, combining performance indicators, science mapping, and a focused full-text synthesis of highly cited papers. The literature grows sharply after 2019, peaks in 2025, and shows geographically uneven production, with collaboration structured around a small set of hubs. The thematic structure suggests that, during the pandemic era, infodemic-related research served as a catalyst, intensifying scientific attention to fake news and disinformation and expanding the associated detection and monitoring agendas. In addition, socio-political harm constructs such as hate speech, extremism, and polarization appear as recurrent and structurally central targets, highlighting that election-relevant work often extends beyond veracity assessment toward monitoring discourse risks. Blockchain also emerges as a novel and adjacent integrity theme, aligned with authenticity and provenance-oriented mitigation rather than mainstream detection pipelines. AI for electoral disinformation is not reducible to veracity classification, as influential studies also target automation and coordinated behavior, verification support, diffusion analysis, and estimation frameworks that focus on exposure and impact. Evaluation remains heterogeneous and is often shaped by benchmark settings, making high accuracy values hard to compare and potentially misleading when labeling quality, topic leakage, or context shift are not characterized. Overall, the findings motivate evaluation protocols that align operational objectives with modeling roles and explicitly address robustness to temporal and platform changes, asymmetric error costs during election windows, and representativeness across electoral contexts and languages, while also guiding future work on emerging integrity challenges and governance-relevant deployment settings.

    Original languageAmerican English
    Article number292
    JournalInformation (Switzerland)
    Volume17
    Issue number3
    DOIs
    StateIndexed - Mar 2026

    Bibliographical note

    Publisher Copyright:
    © 2026 by the authors.

    UN SDGs

    This output contributes to the following UN Sustainable Development Goals (SDGs)

    1. SDG 16 - Peace, Justice and Strong Institutions
      SDG 16 Peace, Justice and Strong Institutions

    Keywords

    • electoral disinformation
    • information integrity
    • machine learning
    • platform governance
    • science mapping
    • social media analytics

    Fingerprint

    Dive into the research topics of 'Artificial Intelligence for Detecting Electoral Disinformation on Social Media: Models, Datasets, and Evaluation'. Together they form a unique fingerprint.

    Cite this