Graph convolutional networks (GCNs) have obtained remarkable performance in skeleton-based action recognition. However, previous approaches fail to capture the implicit correlations between joints and handle actions across varying time intervals. To address these problems, we propose an adaptive multi-scale difference graph convolution Network (AMD-GCN), which comprises an adaptive spatial graph convolution module (ASGC) and a multi-scale temporal difference convolution module (MTDC). The first module is capable of acquiring data-dependent and channel-wise graphs that are adaptable to both samples and channels. The second module employs the multi-scale approach to model temporal information across a range of time scales. Additionally, the MTDC incorporates an attention-enhanced module and difference convolution to accentuate significant channels and enhance temporal features, respectively. Finally, we propose a multi-stream framework for integrating diverse skeletal modalities to achieve superior performance. Our AMD-GCN approach was extensively tested and proven to outperform the current state-of-the-art methods on three widely recognized benchmarks: the NTU-RGB+D, NTU-RGB+D 120, and Kinetics Skeleton datasets.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.