Projects per year
Abstract
Event cameras asynchronously capture pixel-level intensity changes in scenes and output a stream of events. Compared with traditional frame-based cameras, they can offer competitive imaging characteristics: low latency, high dynamic range, and low power consumption. It means that event cameras are ideal for vision tasks in dynamic scenarios, such as human action recognition. The best-performing event-based algorithms convert events into frame-based representations and feed them into existing learning models. However, generating informative frames for long-duration event streams is still a challenge since event cameras work asynchronously without a fixed frame rate. In this work, we propose a novel frame-based representation named Compact Event Image (CEI) for action recognition. This representation is generated by a self-attention based module named Event Tubelet Compressor (EVTC) in a learnable way. The EVTC module adaptively summarizes the long-term dynamics and temporal patterns of events into a CEI frame set. We can combine EVTC with conventional video backbones for end-to-end event-based action recognition. We evaluate our approach on three benchmark datasets, and experimental results show it outperforms state-of-the-art methods by a large margin. ©2022 IEEE.
| Original language | English |
|---|---|
| Title of host publication | 2022 7th International Conference on Control, Robotics and Cybernetics (CRC 2022) |
| Publisher | IEEE |
| Pages | 12-16 |
| ISBN (Electronic) | 978-1-6654-7306-4, 978-1-6654-7305-7 |
| ISBN (Print) | 978-1-6654-7307-1 |
| DOIs | |
| Publication status | Published - Dec 2022 |
| Event | 7th International Conference on Control, Robotics and Cybernetics (CRC 2022) - Virtual, Zhanjiang, China Duration: 15 Dec 2022 → 17 Dec 2022 http://www.iccrc.org/ |
Conference
| Conference | 7th International Conference on Control, Robotics and Cybernetics (CRC 2022) |
|---|---|
| Place | China |
| City | Zhanjiang |
| Period | 15/12/22 → 17/12/22 |
| Internet address |
Funding
This work was supported in part by the Research Grants Council of Hong Kong under Project CityU 11213420, in part by the National Natural Science Foundation of China under Grants 61873220, 62173286 and 62203024, and in part by the Science and Technology Development Fund, Macau, under Grant 0022/2019/AKP.
Research Keywords
- Event camera
- Representation learning
- Self-attention mechanism
- Human action recognition
Fingerprint
Dive into the research topics of 'Event Tubelet Compressor: Generating Compact Representations for Event-Based Action Recognition'. Together they form a unique fingerprint.Projects
- 1 Finished
-
GRF: Gaze Tracking and its Integration with Human-Robot Cooperation
LI, Y. F. (Principal Investigator / Project Coordinator) & CHEN, H. (Co-Investigator)
1/01/21 → 24/06/25
Project: Research