Abstract
While interpretability tools are intended to help people better understand machine learning (ML), we find that they can, in fact, impair understanding. This paper presents a pre-registered, controlled experiment showing that ML practitioners (N = 119) spent 5x less time on task, and were 17% less accurate about the data and model, when given access to interpretability tools. We present bounded rationality as the theoretical reason behind these findings. Bounded rationality presumes human departures from perfect rationality, and it is often effectuated by satisficing, i.e., an inclination towards “good enough” understanding. Adding interactive elements-a strategy often employed to promote deliberative thinking and engagement, and tested in our experiment-also does not help. We discuss implications for interpretability designers and researchers related to how cognitive and contextual factors can affect the effectiveness of interpretability tool use.
Original language | English (US) |
---|---|
Article number | 77 |
Journal | Proceedings of the ACM on Human-Computer Interaction |
Volume | 8 |
Issue number | 1 |
DOIs | |
State | Published - Apr 23 2024 |
Bibliographical note
Publisher Copyright:© 2024 Copyright held by the owner/author(s). Publication rights licensed to ACM.
Keywords
- bounded rationality
- cognitive science
- explainability
- interpretability