Skip to main navigation Skip to search Skip to main content

Can We Mitigate Backdoor Attack Using Adversarial Detection Methods?

Kaidi Jin, Tianwei Zhang, Chao Shen*, Yufei Chen, Ming Fan, Chenhao Lin, Ting Liu

*Corresponding author for this work

Research output: Journal Publications and ReviewsRGC 21 - Publication in refereed journalpeer-review

Abstract

Deep Neural Networks are well known to be vulnerable to adversarial attacks and backdoor attacks, where minor modifications on the input are able to mislead the models to give wrong results. Although defenses against adversarial attacks have been widely studied, investigation on mitigating backdoor attacks is still at an early stage. It is unknown whether there are any connections and common characteristics between the defenses against these two attacks. We conduct comprehensive studies on the connections between adversarial examples and backdoor examples of Deep Neural Networks to seek to answer the question: can we detect backdoor using adversarial detection methods. Our insights are based on the observation that both adversarial examples and backdoor examples have anomalies during the inference process, highly distinguishable from benign samples. As a result, we revise four existing adversarial defense methods for detecting backdoor examples. Extensive evaluations indicate that these approaches provide reliable protection against backdoor attacks, with a higher accuracy than detecting adversarial examples. These solutions also reveal the relations of adversarial examples, backdoor examples and normal samples in model sensitivity, activation space and feature space. This is able to enhance our understanding about the inherent features of these two attacks and the defense opportunities.

© 2022 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission.
Original languageEnglish
Pages (from-to)2867-2881
JournalIEEE Transactions on Dependable and Secure Computing
Volume20
Issue number4
Online published28 Jul 2022
DOIs
Publication statusPublished - Jul 2023

Research Keywords

  • Adversarial attacks
  • backdoor attacks
  • deep neural networks
  • robustness

Fingerprint

Dive into the research topics of 'Can We Mitigate Backdoor Attack Using Adversarial Detection Methods?'. Together they form a unique fingerprint.

Cite this