2020
DOI: 10.1007/978-3-030-58610-2_16
|View full text |Cite
|
Sign up to set email alerts
|

Edge-Aware Graph Representation Learning and Reasoning for Face Parsing

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
33
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
3
3
2

Relationship

0
8

Authors

Journals

citations
Cited by 49 publications
(33 citation statements)
references
References 27 publications
0
33
0
Order By: Relevance
“…To explore high-order relations between the lower-level local features from CIM and higher-level cues from CFM. We introduce the non-local [77], [78] operation under graph convolution domain [79] to implement our similarity aggregation module (SAM). As a result, SAM can inject detailed appearance features into high-level semantic features using global attention.…”
Section: E Similarity Aggregation Modulementioning
confidence: 99%
See 2 more Smart Citations
“…To explore high-order relations between the lower-level local features from CIM and higher-level cues from CFM. We introduce the non-local [77], [78] operation under graph convolution domain [79] to implement our similarity aggregation module (SAM). As a result, SAM can inject detailed appearance features into high-level semantic features using global attention.…”
Section: E Similarity Aggregation Modulementioning
confidence: 99%
“…K T is the transpose of K and f is the correlation attention map. After obtaining the correlation attention map f , we multiply it with the feature map Q, and the result features are fed to the graph convolutional layer [78] GCN(•), leading to G ∈ R 4×4×16 . Same to [78], we calculate the inner product between f and G as Eqn.…”
Section: E Similarity Aggregation Modulementioning
confidence: 99%
See 1 more Smart Citation
“…We perform face landmark detection on the videos before training. We optionally use a state-ofthe-art face parsing algorithm [44] to obtain parsing maps of the videos. During training, we randomly select pairs of source and driving images from each training video.…”
Section: Training Lossesmentioning
confidence: 99%
“…Recently, graph convolution [11] has been incorporated into computer vision tasks for globally reasoning, which can be generally summarized as two kinds of approaches: feature space graph convolution and coordinate space graph convolution. The feature space graph convolution captures interdependencies along the channel dimensions of the feature map, which projects the feature into a non-coordinate space [12][13][14][15]; whistle coordinate space graph convolution explicitly models the spatial relationships between pixels [16][17][18][19][20], which projects the feature into a new coordinate space, to produce coherent prediction between the disjoint infections.…”
Section: Introductionmentioning
confidence: 99%