Evaluating Human-Centered AI Explanations: Introduction of an XAI Evaluation Framework for Fact-Checking

Vera Schmitt, Balázs Patrik Csomor, Joachim Meyer, Luis Felipe Villa-Areas, Charlott Jakob, Tim Polzehl, Sebastian Möller

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

3 Scopus citations

Abstract

The rapidly increasing amount of online information and the advent of Generative Artificial Intelligence (GenAI) make the manual verification of information impractical. Consequently, AI systems are deployed to detect disinformation and deepfakes. Prior studies have indicated that combining AI and human capabilities yields enhanced performance in detecting disinformation. Furthermore, the European Union (EU) AI Act mandates human supervision for AI applications in areas impacting essential human rights, like freedom of speech, necessitating that AI systems be transparent and provide adequate explanations to ensure comprehensibility. Extensive research has been conducted on incorporating explainability (XAI) attributes to augment AI transparency, yet these often miss a human-centric assessment. The effectiveness of such explanations also varies with the user's prior knowledge and personal attributes. Therefore, we developed a framework for validating XAI features for the collaborative human-AI fact-checking task. The framework allows the testing of XAI features with objective and subjective evaluation dimensions and follows human-centric design principles when displaying information about the AI system to the users. The framework was tested in a crowdsourcing experiment with 433 participants, including 406 crowdworkers and 27 journalists for the collaborative disinformation detection task. The tested XAI features increase the AI system's perceived usefulness, understandability, and trust. With this publication, the XAI evaluation framework is made open source.

Original languageEnglish
Title of host publicationMAD 2024 - Proceedings of the 3rd ACM International Workshop on Multimedia AI against Disinformation
PublisherAssociation for Computing Machinery
Pages91-100
Number of pages10
ISBN (Electronic)9798400705526
DOIs
StatePublished - 10 Jun 2024
Event3rd ACM International Workshop on Multimedia AI against Disinformation, MAD 2024 - Phuket, Thailand
Duration: 10 Jun 2024 → …

Publication series

NameACM International Conference Proceeding Series

Conference

Conference3rd ACM International Workshop on Multimedia AI against Disinformation, MAD 2024
Country/TerritoryThailand
CityPhuket
Period10/06/24 → …

Funding

FundersFunder number
Bundesministerium für Bildung und Forschung03RU2U151C

    Keywords

    • Human-centered eXplanations
    • blind trust in AI systems
    • objective and subjective evaluation of eXplanations

    Fingerprint

    Dive into the research topics of 'Evaluating Human-Centered AI Explanations: Introduction of an XAI Evaluation Framework for Fact-Checking'. Together they form a unique fingerprint.

    Cite this