Abstract

As deepfake information manipulation technology continues to evolve and propagate, its potential to mislead the public poses growing threat to societal trust. This paper outlines our research agenda exploring the role of explainable AI (XAI) in cyber defense and its effectiveness in safeguarding reality. Our study examines mechanisms for unveiling deepfakes in ways that enhance sensemaking and strengthen individual cyber defense self-efficacy in distinguishing authentic from manipulated information. To achieve this, we designed and simulated human-AI collaboration experiments with participants from the United States and Italy in Spring 2025. These experiments will generate paired datasets of real and deepfake artifacts across audio, graphic, visual and textual content. XAI—defined by the completeness and relevance of explanations regarding deepfake information—will be modeled based on the insights from the collaboration. Ultimately, this study contributes to social cybersecurity by empowering individuals and communities to recognize and defend against deepfake information manipulation.

Share

COinS