Motif-GCNs with Local and Non-Local Temporal Blocks for Skeleton-Based Action Recognition
Research output: Journal Publications and Reviews (RGC: 21, 22, 62) › 21_Publication in refereed journal › peer-review
Author(s)
Related Research Unit(s)
Detail(s)
Original language | English |
---|---|
Pages (from-to) | 2009-2023 |
Journal / Publication | IEEE Transactions on Pattern Analysis and Machine Intelligence |
Volume | 45 |
Issue number | 2 |
Online published | 26 Apr 2022 |
Publication status | Published - Feb 2023 |
Link(s)
Abstract
Recent works have achieved remarkable performance for action recognition with human skeletal data by utilizing graph convolutional models. Existing models mainly focus on developing graph convolutions to encode structural properties of the skeletal graph. Some recent works further take sample-dependent relationships among joints into consideration. However, the complex relationships are difficult to learn. In this paper, we propose a motif-based graph convolution method, which makes use of sample-dependent latent relations among non-physically connected joints to impose a high-order locality and assigns different semantic roles to physical neighbors of a joint to encode hierarchical structures. Furthermore, we propose a sparsity-promoting loss function to learn a sparse motif adjacency matrix for latent dependencies in non-physical connections. For extracting effective temporal information, we propose an efficient local temporal block. It adopts partial dense connections to reuse temporal features in local time windows, and enrich a variety of information flow by gradient combination. In addition, we introduce a non-local temporal block to capture global dependencies among frames. Comprehensive experiments on four large-scale datasets show that our model outperforms the state-of-the-art methods. Our code is publicly available at https://github.com/wenyh1616/SAMotif-GCN.
Research Area(s)
- Action Recognition, Convolutional codes, Feature extraction, Graph Convolutional Neural Networks, Joints, Non-Local Block, Skeleton, Skeleton Sequence, Sparse matrices, Spatio-Temporal Attention, Topology, Training
Citation Format(s)
Motif-GCNs with Local and Non-Local Temporal Blocks for Skeleton-Based Action Recognition. / Wen, Yu-Hui; Gao, Lin; Fu, Hongbo et al.
In: IEEE Transactions on Pattern Analysis and Machine Intelligence, Vol. 45, No. 2, 02.2023, p. 2009-2023.
In: IEEE Transactions on Pattern Analysis and Machine Intelligence, Vol. 45, No. 2, 02.2023, p. 2009-2023.
Research output: Journal Publications and Reviews (RGC: 21, 22, 62) › 21_Publication in refereed journal › peer-review