2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW) 2023
DOI: 10.1109/cvprw59228.2023.00386
|View full text |Cite
|
Sign up to set email alerts
|

Text2Concept: Concept Activation Vectors Directly from Text

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Publication Types

Select...
1
1

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(2 citation statements)
references
References 14 publications
0
2
0
Order By: Relevance
“…Moreover, recent advancements have seen the introduction of invertible CAVs to interpret vision models with non-negative CAVs [78], and Text2Concept which extends the CAV framework to NLP, allowing the extraction of interpretable vectors from text [79]. Concept Activation Regions (CARs) further generalize this framework by using a collection of CAVs to define the decision boundaries in DNN models [80].…”
Section: Concept Activation Vectors and Derivativesmentioning
confidence: 99%
“…Moreover, recent advancements have seen the introduction of invertible CAVs to interpret vision models with non-negative CAVs [78], and Text2Concept which extends the CAV framework to NLP, allowing the extraction of interpretable vectors from text [79]. Concept Activation Regions (CARs) further generalize this framework by using a collection of CAVs to define the decision boundaries in DNN models [80].…”
Section: Concept Activation Vectors and Derivativesmentioning
confidence: 99%
“…In the past few years, image-text contrastively pre-trained multimodal models such as CLIP (Radford et al 2021a) have shown tremendous ability to perform zero-shot classification (Mu et al 2021;Minderer et al 2022), imagetext retrieval (Diwan et al 2022;Thrush et al 2022) and image-captioning (Yu et al 2022;Li et al 2022;Mokady, Hertz, and Bermano 2021). These contrastive models are also used as a part of various state-of-the-art pipelines for downstream tasks such as segmentation (Wang et al 2021;Lüddecke and Ecker 2021), object-detection (Minderer et al 2022;Zhong et al 2021) and model interpretability (Moayeri et al 2023). However, recent works have shown that these models fail on visio-linguistic reasoning tasks, for example identifying the relative position between objects in an image.…”
Section: Introductionmentioning
confidence: 99%