On Certifying Non-uniform Bound against Adversarial Attacks

Chen Liu*, Ryota Tomioka, Volkan Cevher

*Corresponding author for this work

Research output: Chapters, Conference Papers, Creative and Literary WorksRGC 32 - Refereed conference paper (with host publication)peer-review

13 Citations (Scopus)

Abstract

This work studies the robustness certification problem of neural network models, which aims to find certified adversary-free regions as large as possible around data points. In contrast to the existing approaches that seek regions bounded uniformly along all input features, we consider non-uniform bounds and use it to study the decision boundary of neural network models. We formulate our target as an optimization problem with nonlinear constraints. Then, a framework applicable for general feedforward neural networks is proposed to bound the output logits so that the relaxed problem can be solved by the augmented Lagrangian method. Our experiments show the non-uniform bounds have larger volumes than uniform ones. Compared with normal models, the robust models have even larger non-uniform bounds and better interpretability. Further, the geometric similarity of the non-uniform bounds gives a quantitative, data-agnostic metric of input features’ robustness.
Original languageEnglish
Title of host publicationProceedings of the 36th International Conference on Machine Learning, ICML 2019
EditorsKamalika Chaudhuri, Ruslan Salakhutdinov
PublisherInternational Conference on Machine Learning (ICML)
Pages4072-4081
Publication statusPublished - Jun 2019
Externally publishedYes
Event36th International Conference on Machine Learning (ICML 2019) - Long Beach, United States
Duration: 9 Jun 201915 Jun 2019
https://icml.cc/

Publication series

NameProceedings of Machine Learning Research
Volume97
ISSN (Print)2640-3498

Conference

Conference36th International Conference on Machine Learning (ICML 2019)
PlaceUnited States
CityLong Beach
Period9/06/1915/06/19
Internet address

Fingerprint

Dive into the research topics of 'On Certifying Non-uniform Bound against Adversarial Attacks'. Together they form a unique fingerprint.

Cite this