papers AI Learner
The Github is limit! Click to go to the new site.

On The Stability of Interpretable Models

2019-03-15
Riccardo Guidotti, Salvatore Ruggieri

Abstract

Interpretable classification models are built with the purpose of providing a comprehensible description of the decision logic to an external oversight agent. When considered in isolation, a decision tree, a set of classification rules, or a linear model, are widely recognized as human-interpretable. However, such models are generated as part of a larger analytical process. Bias in data collection and preparation, or in model’s construction may severely affect the accountability of the design process. We conduct an experimental study of the stability of interpretable models with respect to feature selection, instance selection, and model selection. Our conclusions should raise awareness and attention of the scientific community on the need of a stability impact assessment of interpretable models.

Abstract (translated by Google)
URL

http://arxiv.org/abs/1810.09352

PDF

http://arxiv.org/pdf/1810.09352


Similar Posts

Comments