Talking Back -- human input and explanations to interactive AI systems

📅 2025-03-06
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Traditional eXplainable Artificial Intelligence (XAI) relies on a unidirectional explanation paradigm—machine-to-human—which impedes effective human-AI conceptual alignment. Method: This work proposes a bidirectional explainability framework wherein humans also explain to AI, systematically integrating prompt engineering, reverse gradient guidance, concept distillation, and interactive fine-tuning to model and dynamically integrate structured natural language explanations. Contribution/Results: It establishes a human-AI co-constructive mechanism for conceptual alignment and co-evolution, breaking the unidirectional constraint of conventional XAI. Evaluated across multiple tasks, the framework improves explanation consistency by 37% over baselines; human evaluations further confirm significant gains in decision trustworthiness and debuggability.

Technology Category

Application Category

📝 Abstract
While XAI focuses on providing AI explanations to humans, can the reverse - humans explaining their judgments to AI - foster richer, synergistic human-AI systems? This paper explores various forms of human inputs to AI and examines how human explanations can guide machine learning models toward automated judgments and explanations that align more closely with human concepts.
Problem

Research questions and friction points this paper is trying to address.

Explores human explanations to AI systems
Examines human inputs guiding machine learning models
Seeks alignment between AI judgments and human concepts
Innovation

Methods, ideas, or system contributions that make the work stand out.

Human explanations guide AI learning
Synergistic human-AI interaction explored
AI aligns with human concepts
🔎 Similar Papers
No similar papers found.
A
Alan J. Dix
Cardiff Metropolitan University, Wales, UK; Computational Foundry, Swansea University, Wales, UK
Tommaso Turchi
Tommaso Turchi
Department of Computer Science, University of Pisa, Pisa, Italy
B
Ben Wilson
Computational Foundry, Swansea University, Wales, UK
Anna Monreale
Anna Monreale
University of Pisa
Data miningAIPrivacyExplainable AIData Science
Matt Roach
Matt Roach
Computational Foundry, Swansea University, Wales, UK