We introduce a method capable of answering the second question both directly and efficiently. In this work, we limit the inputs to be images. In general, the proposed method generates explanations in the input space of any model capable of efficient evaluation and gradient evaluation. It does not require any knowledge of the underlying classifier nor use heuristics in its explanation generation, and it is computationally fast to evaluate. We provide extensive experimental results on three different datasets, showing the robustness of our approach, and its superiority for gaining insight into the inner representations of machine learning models. As an example, we demonstrate our method can detect and explain how a network trained to recognize hair color actually detects eye color, whereas other methods cannot find this bias in the trained classifier.
Amir Feghahati, Christian R. Shelton, Michael J. Pazzani, and Kevin Tang (2020). "CDeepEx: Contrastive Deep Explanations." European Conference on Artificial Intelligence. |
@inproceedings{Fegetal20, author = "Amir Feghahati and Christian R. Shelton and Michael J. Pazzani and Kevin Tang", title = "{CDeepEx}: Contrastive Deep Explanations", booktitle = "European Conference on Artificial Intelligence", booktitleabbr = "ECAI", year = 2020, }