VMV-GCN: Volumetric Multi-View Based Graph CNN for Event Stream Classification

Bochen Xie, Yongjian Deng, Zhanpeng Shao, Hai Liu, Youfu Li*

*Corresponding author for this work

Research output: Journal Publications and ReviewsRGC 21 - Publication in refereed journalpeer-review

41 Citations (Scopus)

Abstract

Event cameras can perceive pixel-level brightness changes to output asynchronous event streams, and have notable advantages in high temporal resolution, high dynamic range and low power consumption for challenging vision tasks. To apply existing learning methods on event data, many researchers integrate sparse events into dense frame-based representations which can work with convolutional neural networks directly. Although these works achieve high performance on event-based classification, their models need lots of parameters to process dense event frames which do not fit with the sparsity of event data. To utilize the sparse nature of events, we propose a voxel-wise graph learning model (VMV-GCN) for spatio-temporal feature learning on event streams. Specifically, we design the volumetric multi-view fusion module (VMVF) to extract spatial and temporal information from views of voxelized event data. Then we take representative event voxels as vertices and use a novel dual-graph construction strategy to connect them. By aggregating information based on relationships of vertices, the proposed dynamic neighborhood feature learning module (DNFL) can capture discriminative spatio-temporal features on dynamically updated graphs. Experiments show that our method achieves state-of-the-art performance with low model complexity on event-based classification tasks, such as object classification and action recognition. © 2022 IEEE.
Original languageEnglish
Pages (from-to)1976-1983
JournalIEEE Robotics and Automation Letters
Volume7
Issue number2
Online published6 Jan 2022
DOIs
Publication statusPublished - Apr 2022

Funding

This work was supported in part by the Research Grants Council of Hong Kong under Project CityU11213420 and in part by the National Natural Science Foundation of China under Grants 61873220, 62173286, and 62177018. This letter was recommended for publication by Associate Editor G. Gallego and Editor C. Cadena Lerma upon evaluation of the reviewers’ comments

Research Keywords

  • Brightness
  • Cameras
  • Complexity theory
  • Deep learning for visual perception
  • object detection
  • segmentation and categorization

RGC Funding Information

  • RGC-funded

Fingerprint

Dive into the research topics of 'VMV-GCN: Volumetric Multi-View Based Graph CNN for Event Stream Classification'. Together they form a unique fingerprint.

Cite this