Interpretability Gone Bad: The Role of Bounded Rationality in How Practitioners Understand Machine Learning

Harmanpreet Kaur, Matthew R. Conrad, Davis Rule, Cliff Lampe, Eric Gilbert

Research output: Contribution to journalArticlepeer-review

Abstract

While interpretability tools are intended to help people better understand machine learning (ML), we find that they can, in fact, impair understanding. This paper presents a pre-registered, controlled experiment showing that ML practitioners (N = 119) spent 5x less time on task, and were 17% less accurate about the data and model, when given access to interpretability tools. We present bounded rationality as the theoretical reason behind these findings. Bounded rationality presumes human departures from perfect rationality, and it is often effectuated by satisficing, i.e., an inclination towards “good enough” understanding. Adding interactive elements-a strategy often employed to promote deliberative thinking and engagement, and tested in our experiment-also does not help. We discuss implications for interpretability designers and researchers related to how cognitive and contextual factors can affect the effectiveness of interpretability tool use.

Original languageEnglish (US)
Article number77
JournalProceedings of the ACM on Human-Computer Interaction
Volume8
Issue number1
DOIs
StatePublished - Apr 23 2024

Bibliographical note

Publisher Copyright:
© 2024 Copyright held by the owner/author(s). Publication rights licensed to ACM.

Keywords

  • bounded rationality
  • cognitive science
  • explainability
  • interpretability

Fingerprint

Dive into the research topics of 'Interpretability Gone Bad: The Role of Bounded Rationality in How Practitioners Understand Machine Learning'. Together they form a unique fingerprint.

Cite this