Explaining with Counter Visual Attributes and Examples
| Authors | |
|---|---|
| Publication date | 2020 |
| Book title | ICMR '20 |
| Book subtitle | proceedings of the 2020 International Conference on Multimedia Retrieval : June 08-11, 2020, Dublin, Ireland |
| ISBN (electronic) |
|
| Event | 10th ACM International Conference on Multimedia Retrieval, ICMR 2020 |
| Pages (from-to) | 35-43 |
| Publisher | New York, NY: The Association for Computing Machinery |
| Organisations |
|
| Abstract |
In this paper, we aim to explain the decisions of neural networks by utilizing multimodal information. That is counter-intuitive attributes and counter visual examples which appear when perturbed samples are introduced. Different from previous work on interpreting decisions using saliency maps, text, or visual patches we propose to use attributes and counter-attributes, and examples and counter-examples as part of the visual explanations. When humans explain visual decisions they tend to do so by providing attributes and examples. Hence, inspired by the way of human explanations in this paper we provide attribute-based and example-based explanations. Moreover, humans also tend to explain their visual decisions by adding counter-attributes and counter-examples to explain what isnot seen. We introduce directed perturbations in the examples to observe which attribute values change when classifying the examples into the counter classes. This delivers intuitive counter-attributes and counter-examples. Our experiments with both coarse and fine-grained datasets show that attributes provide discriminating and human-understandable intuitive and counter-intuitive explanations.
|
| Document type | Conference contribution |
| Language | English |
| Published at | https://doi.org/10.1145/3372278.3390672 |
| Other links | https://ivi.fnwi.uva.nl/vislab/publication/gulshad-icmr-2020/ |
| Permalink to this page | |