Abstract
Deep convolutional networks have recently gained much attention because of their impressive performance on some visual tasks. However, it is still not clear why they achieve such great success. In this paper, a novel approach called Filter Sensitive Area Generation Network (FSAGN), has been proposed to interpret what the convolutional filters have learnt after training CNNs. Given any trained CNN model, the proposed method aims to figure out which object part each filter represents in a high conv-layer, through appropriate input image mask which filters out unrelated area. In order to obtain such a mask, a mask generation network is designed and the corresponding loss function is defined to evaluate the changes of feature maps before and after mask operation. Experiments on multiple datasets and networks show that FSAGN clarifies the knowledge representations of each filter and how small disturbance on specific object parts affects the performance of CNNs. © 2018, Springer Nature Switzerland AG.
| Original language | English |
|---|---|
| Title of host publication | Neural Information Processing - 25th International Conference, ICONIP 2018, Proceedings |
| Publisher | Springer Verlag |
| Pages | 192-203 |
| Volume | 11301 LNCS |
| ISBN (Print) | 9783030041663 |
| DOIs | |
| Publication status | Published - 2018 |
| Externally published | Yes |
| Event | 25th International Conference on Neural Information Processing, ICONIP 2018 - Siem Reap, Cambodia Duration: 13 Dec 2018 → 16 Dec 2018 |
Publication series
| Name | Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) |
|---|---|
| Volume | 11301 LNCS |
| ISSN (Print) | 0302-9743 |
| ISSN (Electronic) | 1611-3349 |
Conference
| Conference | 25th International Conference on Neural Information Processing, ICONIP 2018 |
|---|---|
| Place | Cambodia |
| City | Siem Reap |
| Period | 13/12/18 → 16/12/18 |
Bibliographical note
Publication details (e.g. title, author(s), publication statuses and dates) are captured on an “AS IS” and “AS AVAILABLE” basis at the time of record harvesting from the data source. Suggestions for further amendments or supplementary information can be sent to <a href="mailto:[email protected]">[email protected]</a>.Funding
This work was supported in part by the National Key Research and Development Program of China (2017YFB1300203), in part by the National Natural Science Foundation of China under Grant 91648205.
Research Keywords
- Convolutional neural network
- Interpretability
- Knowledge representations