Forecasting Semantic Bird-Eye-View Maps for Autonomous Driving

Shuang Gao*, Qiang Wang, David Navarro-Alarcon, Yuxiang Sun

*Corresponding author for this work

Research output: Chapters, Conference Papers, Creative and Literary WorksRGC 32 - Refereed conference paper (with host publication)peer-review

3 Citations (Scopus)

Abstract

Correctly understanding surrounding environments is a fundamental capability for autonomous driving. Semantic forecasting of bird-eye-view (BEV) maps can provide semantic perception information in advance, which is important for environment understanding. Currently, the research works on combining semantic forecasting and semantic BEV map generation is limited. Most existing work focuses on individual tasks only. In this work, we attempt to forecast semantic BEV maps in an end-to-end framework for future front-view (FV) images. To this end, we predict depth distributions and context features for FV input images and then forecast depth-context features for the future. The depth-context features are finally converted to the future semantic BEV maps. We conduct ablation studies and create baselines for evaluation and comparison. The results demonstrate that our network achieves superior performance. © 2024 IEEE.
Original languageEnglish
Title of host publication2024 IEEE Intelligent Vehicles Symposium (IV)
PublisherIEEE
Pages509-514
ISBN (Electronic)9798350348811
ISBN (Print)979-8-3503-4882-8
DOIs
Publication statusPublished - 2024
Event35th IEEE Intelligent Vehicles Symposium (IV 2024) - Landing Convention Center, Jeju Island, Korea, Republic of
Duration: 2 Jun 20245 Jun 2024
https://ieee-iv.org/2024/

Publication series

NameIEEE Intelligent Vehicles Symposium, Proceedings
ISSN (Print)1931-0587
ISSN (Electronic)2642-7214

Conference

Conference35th IEEE Intelligent Vehicles Symposium (IV 2024)
PlaceKorea, Republic of
CityJeju Island
Period2/06/245/06/24
Internet address

Bibliographical note

Full text of this publication does not contain sufficient affiliation information. With consent from the author(s) concerned, the Research Unit(s) information for this record is based on the existing academic department affiliation of the author(s).

Funding

This work was supported in part by Guangdong Basic and Applied Basic Research Foundation under Grant 2022A1515010116, in part by Hong Kong Research Grants Council under Grant 15222523, and in part by City University of Hong Kong under Grant 9610675.

Fingerprint

Dive into the research topics of 'Forecasting Semantic Bird-Eye-View Maps for Autonomous Driving'. Together they form a unique fingerprint.

Cite this