Manipulating and Measuring Model Interpretability

"Machine learning is increasingly used to make decisions that affect people's lives in critical domains like criminal justice, fair lending, and medicine. While most of the research in machine learning focuses on improving the performance of models on held-out datasets, this is seldom enou...

Full description

Bibliographic Details
Main Author: Poursabzi-Sangdeh, Forough
Format: eBook
Language:English
Published: [Place of publication not identified] O'Reilly Media 2019
Subjects:
Online Access:
Collection: O'Reilly - Collection details see MPG.ReNa
Description
Summary:"Machine learning is increasingly used to make decisions that affect people's lives in critical domains like criminal justice, fair lending, and medicine. While most of the research in machine learning focuses on improving the performance of models on held-out datasets, this is seldom enough to convince end users that these models are trustworthy and reliable in the wild. To address this problem, a new line of research has emerged that focuses on developing interpretable machine learning methods and helping end users make informed decisions. Despite the growing body of work in developing interpretable models, there is still no consensus on the definition and quantification of interpretability ... Forough approaches the problem of interpretability from an interdisciplinary perspective built on decades of research in psychology, cognitive science, and social science to understand human behavior and trust. She describes a set of controlled user experiments in which researchers manipulated various design factors in models that are commonly thought to make them more or less interpretable and measured their influence on users' behavior."--Resource description page
Item Description:Title from resource description page (Safari, viewed November 12, 2019)
Physical Description:1 streaming video file (39 min., 42 sec.)