Unsupervised video-based action recognition using two-stream generative adversarial network
Research output: Journal Publications and Reviews › RGC 21 - Publication in refereed journal › peer-review
Author(s)
Related Research Unit(s)
Detail(s)
Original language | English |
---|---|
Pages (from-to) | 5077-5091 |
Number of pages | 15 |
Journal / Publication | Neural Computing and Applications |
Volume | 36 |
Issue number | 9 |
Online published | 26 Dec 2023 |
Publication status | Published - Mar 2024 |
Link(s)
Abstract
Video-based action recognition faces many challenges, such as complex and varied dynamic motion, spatio-temporal similar action factors, and manual labeling of archived videos over large datasets. How to extract discriminative spatio-temporal action features in videos with resisting the effect of similar factors in an unsupervised manner is pivotal. For that, this paper proposes an unsupervised video-based action recognition method, called two-stream generative adversarial network (TS-GAN), which comprehensively learns the static texture and dynamic motion information inherited in videos with taking the detail information and global information into account. Specifically, the extraction of the spatio-temporal information in videos is achieved by a two-stream GAN. Considering that proper attention to detail is capable of alleviating the influence of spatio-temporal similar factors to the network, a global-detailed layer is proposed to resist similar factors via fusing intermediate features (i.e., detailed action information) and high-level semantic features (i.e., global action information). It is worthwhile of mentioning that the proposed TS-GAN does not require complex pretext tasks or the construction of positive and negative sample pairs, compared with recent unsupervised video-based action recognition methods. Extensive experiments conducted on the UCF101 and HMDB51 datasets have demonstrated that the proposed TS-GAN is superior to multiple classical and state-of-the-art unsupervised action recognition methods. © The Author(s), under exclusive licence to Springer-Verlag London Ltd., part of Springer Nature 2023.
Research Area(s)
- Action recognition, Two-stream generative adversarial network, Unsupervised learning
Bibliographic Note
Research Unit(s) information for this publication is provided by the author(s) concerned.
Citation Format(s)
Unsupervised video-based action recognition using two-stream generative adversarial network. / Lin, Wei; Zeng, Huanqiang; Zhu, Jianqing et al.
In: Neural Computing and Applications, Vol. 36, No. 9, 03.2024, p. 5077-5091.
In: Neural Computing and Applications, Vol. 36, No. 9, 03.2024, p. 5077-5091.
Research output: Journal Publications and Reviews › RGC 21 - Publication in refereed journal › peer-review