2022
DOI: 10.1109/access.2022.3206044
|View full text |Cite
|
Sign up to set email alerts
|

Two-Stream Spatial Graphormer Networks for Skeleton-Based Action Recognition

Abstract: In skeleton-based human action recognition, Transformer, which models the correlations between joint pairs in global topology, has achieved remarkable results. However, compared to many researches on changing graph topology learning in GCN, Transformer self-attention ignores the topology of the skeleton graph when capturing the dependencies between joints. To address these problems, we propose a novel two-stream spatial Graphormer network (2s-SGR), which models joint and bone information using self-attention i… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
1
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
1
1

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(7 citation statements)
references
References 36 publications
0
1
0
Order By: Relevance
“…Graphormer was chosen as our model architecture to address the known limitation of the MPNN architecture and because of its recent success across a wide variety of domains. ,− A complete depiction of our model architecture is present in Figure . Transformers are a type of encoder that are successful because they learn contextual relationships between input embeddings which are aggregated to generate a learned description of the input.…”
Section: Methodsmentioning
confidence: 99%
“…Graphormer was chosen as our model architecture to address the known limitation of the MPNN architecture and because of its recent success across a wide variety of domains. ,− A complete depiction of our model architecture is present in Figure . Transformers are a type of encoder that are successful because they learn contextual relationships between input embeddings which are aggregated to generate a learned description of the input.…”
Section: Methodsmentioning
confidence: 99%
“…Graphormer was chosen as our model architecture to address the known limitation of the MPNN architecture and because of its recent success across a wide variety of do-mains. 46,[56][57][58][59] .A complete depiction of our model architecture is present in Figure 1. Transformers are a type of encoder that are successful because they learn contextual relationships between input embeddings which are aggregated to a learned description of the input.…”
Section: Transformersmentioning
confidence: 99%
“…Graphormer was chosen as our model architecture to address the known limitation of the MPNN architecture and because of its recent success across a wide variety of domains. 46,[56][57][58][59] Transformers are type of encoder that are successful because they learn contextual relationships between input embeddings which are aggregated to a learned description of the input. This is achieved using a self-attention mechanism, where the model assigns varying degrees of importance to between different items in the input sequence.…”
Section: Data Processingmentioning
confidence: 99%