2023
DOI: 10.1007/s00530-023-01164-0
|View full text |Cite
|
Sign up to set email alerts
|

Inceptr: micro-expression recognition integrating inception-CBAM and vision transformer

Haoliang Zhou,
Shucheng Huang,
Yuqiao Xu
Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
5
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
7
1
1

Relationship

0
9

Authors

Journals

citations
Cited by 13 publications
(5 citation statements)
references
References 43 publications
0
5
0
Order By: Relevance
“…Through fusing two different semantic features, they obtain hybrid features with local edge information and global appearance information. Zhou et al [20] addressed Transformer to extract subtle facial motion features and to model the global relationship among Patches. They took the rich relationship among different rich relationships within patches to mine discriminative features.…”
Section: Transformer In Fermentioning
confidence: 99%
“…Through fusing two different semantic features, they obtain hybrid features with local edge information and global appearance information. Zhou et al [20] addressed Transformer to extract subtle facial motion features and to model the global relationship among Patches. They took the rich relationship among different rich relationships within patches to mine discriminative features.…”
Section: Transformer In Fermentioning
confidence: 99%
“…Overcoming the limitations of existing vision transformers, this novel approach balances local spatial relationships and global dependencies. Achieving exemplary performance on datasets such as CASME-I, CASME-II, and SAMM, this methodology integrates convolutional layers and transformers for improved spatial information and feature extraction [16] [17].…”
Section: Architectural Innovationsmentioning
confidence: 99%
“…Zhu et al fuse CNN with Transformer, adopting a sparse self-attention mechanism to extract a sparse representation of feature maps [22], which can reduce the network complexity and improve the recognition performance. The model in [23] integrates inception-CBAM with a Vision Transformer for micro-expression recognition. A recent study uses a combination of CNN and Transformer to fuse local and global facial information in facial MEs [24].…”
Section: Related Workmentioning
confidence: 99%