E-Motion: Future Motion Simulation via Event Sequence Diffusion

Song Wu, Zhiyu Zhu, Junhui Hou, Guangming Shi, Jinjian Wu

Research output: Chapters, Conference Papers, Creative and Literary WorksRGC 32 - Refereed conference paper (with host publication)peer-review

Abstract

Forecasting a typical object's future motion is a critical task for interpreting and interacting with dynamic environments in computer vision. Event-based sensors, which could capture changes in the scene with exceptional temporal granularity, may potentially offer a unique opportunity to predict future motion with a level of detail and precision previously unachievable. Inspired by that, we propose to integrate the strong learning capacity of the video diffusion model with the rich motion information of an event camera as a motion simulation framework. Specifically, we initially employ pre-trained stable video diffusion models to adapt the event sequence dataset. This process facilitates the transfer of extensive knowledge from RGB videos to an event-centric domain. Moreover, we introduce an alignment mechanism that utilizes reinforcement learning techniques to enhance the reverse generation trajectory of the diffusion model, ensuring improved performance and accuracy. Through extensive testing and validation, we demonstrate the effectiveness of our method in various complex scenarios, showcasing its potential to revolutionize motion flow prediction in computer vision applications such as autonomous vehicle guidance, robotic navigation, and interactive media. Our findings suggest a promising direction for future research in enhancing the interpretative power and predictive accuracy of computer vision systems. The source code is publicly available at https://github.com/p4r4mount/E-Motion. © 2024 Neural information processing systems foundation. All rights reserved.
Original languageEnglish
Title of host publicationNeurIPS Proceedings
Subtitle of host publicationAdvances in Neural Information Processing Systems 37 (NeurIPS 2024)
EditorsA. Globerson, L. Mackey, D. Belgrave, A. Fan, U. Paquet, J. Tomczak, C. Zhang
Volume37
Publication statusPublished - 2024
Event38th Annual Conference on Neural Information Processing Systems (NeurIPS 2024) - Vancouver Convention Center, Vancouver, Canada
Duration: 10 Dec 202415 Dec 2024
https://neurips.cc/
https://proceedings.neurips.cc/

Publication series

NameAdvances in Neural Information Processing Systems
PublisherNeural information processing systems foundation
Volume37
ISSN (Print)1049-5258

Conference

Conference38th Annual Conference on Neural Information Processing Systems (NeurIPS 2024)
Abbreviated titleNeurIPS 2024
PlaceCanada
CityVancouver
Period10/12/2415/12/24
Internet address

Funding

This work was supported in part by National Key Research and Development Program of China(2023YFA1008500), in part by NSFC Excellent Young Scientists Fund 62422118, and in part by Hong Kong Innovation and Technology Fund ITS/164/23. The first two authors contributed to this paper equally.

Fingerprint

Dive into the research topics of 'E-Motion: Future Motion Simulation via Event Sequence Diffusion'. Together they form a unique fingerprint.

Cite this