Abstract
It is well known that deep neural networks (DNNs) are vulnerable to adversarial attacks, which are implemented by adding crafted perturbations onto benign examples. Min-max robust optimization based adversarial training can provide a notion of security against adversarial attacks. However, adversarial robustness requires a significantly larger capacity of the network than that for the natural training with only benign examples. This paper proposes a framework of concurrent adversarial training and weight pruning that enables model compression while still preserving the adversarial robustness and essentially tackles the dilemma of adversarial training. Furthermore, this work studies two hypotheses about weight pruning in the conventional setting and finds that weight pruning is essential for reducing the network model size in the adversarial setting; training a small model from scratch even with inherited initialization from the large model cannot achieve neither adversarial robustness nor high standard accuracy. Code is available at https://github.com/yeshaokai/Robustness-Aware-Pruning-ADMM. © 2019 IEEE.
| Original language | English |
|---|---|
| Title of host publication | Proceedings - 2019 International Conference on Computer Vision |
| Publisher | IEEE |
| Pages | 111-120 |
| ISBN (Electronic) | 9781728148038 |
| ISBN (Print) | 978-1-7281-4804-5 |
| DOIs | |
| Publication status | Published - Oct 2019 |
| Externally published | Yes |
| Event | 17th IEEE/CVF International Conference on Computer Vision (ICCV 2019) - COEX Convention Center, Seoul, Korea, Republic of Duration: 27 Oct 2019 → 2 Nov 2019 http://iccv2019.thecvf.com/ |
Publication series
| Name | Proceedings of the IEEE International Conference on Computer Vision |
|---|---|
| ISSN (Print) | 1550-5499 |
| ISSN (Electronic) | 2380-7504 |
Conference
| Conference | 17th IEEE/CVF International Conference on Computer Vision (ICCV 2019) |
|---|---|
| Abbreviated title | ICCV19 |
| Place | Korea, Republic of |
| City | Seoul |
| Period | 27/10/19 → 2/11/19 |
| Internet address |
Funding
This work is partly supported by the National Science Foundation CNS-1932351, Institute for Interdisciplinary Information Core Technology (IIISCT) and Zhongguancun Haihua Institute for Frontier Information Technology.
Fingerprint
Dive into the research topics of 'Adversarial robustness vs. model compression, or both?'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver