Exploring foci of:
arXiv (Cornell University)
Assessing the Local Interpretability of Machine Learning Models
February 2019 • Dylan Slack, Sorelle A. Friedler, Carlos Scheidegger, Chitradeep Dutta Roy
The increasing adoption of machine learning tools has led to calls for accountability via model interpretability. But what does it mean for a machine learning model to be interpretable by humans, and how can this be assessed? We focus on two definitions of interpretability that have been introduced in the machine learning literature: simulatability (a user's ability to run a model on a given input) and "what if" local explainability (a user's ability to correctly determine a model's prediction under local changes …
Machine Learning
Artificial Intelligence
Computer Science
Intuition
Performance Indicator
Management
Economics
Operations Management