Abstract
In Autonomous Driving (AD) systems, perception is both security and safety-critical. Among different attacks on AD perception, object-hiding adversarial attack is one of the most critical ones due to the direct impact on safety-critical driving decisions such as collision avoidance. However, all of the prior works on physical object-hiding adversarial attacks only study the security of the AI component alone rather than with the entire AD system pipeline with closed-loop control. This thus inevitably raises a critical research question: can these prior works actually achieve system-level effects (e.g., vehicle collisions, traffic rule violation) under real-world AD settings with closed-loop control?
To answer this critical question, in this work we take the necessary first step by performing the first measurement study on whether and how effective the existing designs can lead to system-level effects. Our early results find that RP2 and FTE, as two representative examples of prior works, cannot achieve any system-level effect in a representative closed-loop AD setup in common STOP sign-controlled road speeds. In the future, we plan to 1) perform a more comprehensive measurement study using both simulated environments and a real vehicle-sized AD R&D chassis; and 2) analyze the measurement study results and explore new attack designs that can better achieve the system-level effect in AD systems. © 2022 Copyright held by the owner/author(s).
To answer this critical question, in this work we take the necessary first step by performing the first measurement study on whether and how effective the existing designs can lead to system-level effects. Our early results find that RP2 and FTE, as two representative examples of prior works, cannot achieve any system-level effect in a representative closed-loop AD setup in common STOP sign-controlled road speeds. In the future, we plan to 1) perform a more comprehensive measurement study using both simulated environments and a real vehicle-sized AD R&D chassis; and 2) analyze the measurement study results and explore new attack designs that can better achieve the system-level effect in AD systems. © 2022 Copyright held by the owner/author(s).
| Original language | English |
|---|---|
| Title of host publication | CCS ’22 |
| Subtitle of host publication | Proceedings of the 2022 ACM SIGSAC Conference on Computer and Communications Security |
| Publisher | Association for Computing Machinery |
| Pages | 3479-3481 |
| Number of pages | 3 |
| ISBN (Print) | 9781450394505 |
| DOIs | |
| Publication status | Published - 2022 |
| Externally published | Yes |
| Event | 28th ACM SIGSAC Conference on Computer and Communications Security (CCS 2022) - Hybrid , Los Angeles, United States Duration: 7 Nov 2022 → 11 Nov 2022 https://www.sigsac.org/ccs/CCS2022/ |
Publication series
| Name | Proceedings of the ACM Conference on Computer and Communications Security |
|---|---|
| ISSN (Print) | 1543-7221 |
Conference
| Conference | 28th ACM SIGSAC Conference on Computer and Communications Security (CCS 2022) |
|---|---|
| Place | United States |
| City | Los Angeles |
| Period | 7/11/22 → 11/11/22 |
| Internet address |
Funding
This research was supported by the NSF under grants CNS-1932464, CNS-1929771, CNS-2145493, and USDOT UTC Grant 69A3552047138.
UN SDGs
This output contributes to the following UN Sustainable Development Goals (SDGs)
-
SDG 3 Good Health and Well-being
Research Keywords
- Autonomous Driving (AD) system security
- Object Hiding Attack
- System-Level Effect
Fingerprint
Dive into the research topics of 'Poster: On the System-Level Effectiveness of Physical Object-Hiding Adversarial Attack in Autonomous Driving'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver