Learning from Explanations and Demonstrations: A Pilot Study

Silvia Tulli, Sebastian Wallkötter, Ana Paiva, Francisco S. Melo, Mohamed Chetouani


Abstract
AI has become prominent in a growing number of systems, and, as a direct consequence, the desire for explainability in such systems has become prominent as well. To build explainable systems, a large portion of existing research uses various kinds of natural language technologies, e.g., text-to-speech mechanisms, or string visualizations. Here, we provide an overview of the challenges associated with natural language explanations by reviewing existing literature. Additionally, we discuss the relationship between explainability and knowledge transfer in reinforcement learning. We argue that explainability methods, in particular methods that model the recipient of an explanation, might help increasing sample efficiency. For this, we present a computational approach to optimize the learner’s performance using explanations of another agent and discuss our results in light of effective natural language explanations for humans.
Anthology ID:
2020.nl4xai-1.13
Volume:
2nd Workshop on Interactive Natural Language Technology for Explainable Artificial Intelligence
Month:
November
Year:
2020
Address:
Dublin, Ireland
Venue:
NL4XAI
SIG:
SIGGEN
Publisher:
Association for Computational Linguistics
Note:
Pages:
61–66
Language:
URL:
https://aclanthology.org/2020.nl4xai-1.13
DOI:
Bibkey:
Cite (ACL):
Silvia Tulli, Sebastian Wallkötter, Ana Paiva, Francisco S. Melo, and Mohamed Chetouani. 2020. Learning from Explanations and Demonstrations: A Pilot Study. In 2nd Workshop on Interactive Natural Language Technology for Explainable Artificial Intelligence, pages 61–66, Dublin, Ireland. Association for Computational Linguistics.
Cite (Informal):
Learning from Explanations and Demonstrations: A Pilot Study (Tulli et al., NL4XAI 2020)
Copy Citation:
PDF:
https://preview.aclanthology.org/paclic-22-ingestion/2020.nl4xai-1.13.pdf