2022
DOI: 10.1109/tmm.2021.3104411
|View full text |Cite
|
Sign up to set email alerts
|

Relation-Aware Compositional Zero-Shot Learning for Attribute-Object Pair Recognition

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
5
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
6
2

Relationship

0
8

Authors

Journals

citations
Cited by 14 publications
(5 citation statements)
references
References 74 publications
0
5
0
Order By: Relevance
“…Karthik et al [21] further improve simple primitives by estimating the feasibility of each composition through external knowledge to eliminate impossible state-object pairs. Different from [21], [40], our work estimates feasibility with attention mechanism and implies feasibility in the form of probability to enhance simple primitive learning.…”
Section: Open-world Compositional Zero-shot Learningmentioning
confidence: 99%
See 1 more Smart Citation
“…Karthik et al [21] further improve simple primitives by estimating the feasibility of each composition through external knowledge to eliminate impossible state-object pairs. Different from [21], [40], our work estimates feasibility with attention mechanism and implies feasibility in the form of probability to enhance simple primitive learning.…”
Section: Open-world Compositional Zero-shot Learningmentioning
confidence: 99%
“…For example, Li et al [19] mask impossible compositions by computing pair probability based on the distance between state and object categories. Xu et al [40] adopt a keyquery-based attention mechanism to capture the correlation between primitive concepts in a graph to pass messages selectively. In OW-CZSL, similar to [19], Mancini et al [22], [29] propose to utilize the graph structure to model the dependence between state, object, and compositions.…”
Section: Open-world Compositional Zero-shot Learningmentioning
confidence: 99%
“…Li et al [15] disentangle attributes and objects with reversed attention. The second strategy directly predicts compositions by aligning images and textual labels in a shared space and searching for most similar compositions [13,26,35,42]. For example, Nagarajan et al [26] build a composition space by simulating all the visual changes of attributes performed on objects.…”
Section: Related Workmentioning
confidence: 99%
“…The major challenge behind the CZSL lies in how to model the interactions between state and object primitives and extrapolate seen compositions to unseen ones. Existing methods mainly focus on learning a shared embedding space for object-state compositions (Li et al 2020;Naeem et al 2021;Nagarajan and Grauman 2018;Khan et al 2023) or compositional attribute and object classifiers (Purushwalkam et al 2019;Misra, Gupta, and Hebert 2017;Xu et al 2022;Yang et al 2022).…”
Section: Introductionmentioning
confidence: 99%