2022
DOI: 10.48550/arxiv.2204.08261
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Visio-Linguistic Brain Encoding

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 0 publications
0
1
0
Order By: Relevance
“…Further exploration shows that the V2 and V3 are better predicted by multimodal models, with higher layers in image transformers (i.e. unimodal) correlating with late visual areas and vice versa (Oota et al, 2022). It is also suggested that CLIP’s learning process may capture how abstract concepts penetrate early vision in a top-down manner.…”
Section: Introductionmentioning
confidence: 99%
“…Further exploration shows that the V2 and V3 are better predicted by multimodal models, with higher layers in image transformers (i.e. unimodal) correlating with late visual areas and vice versa (Oota et al, 2022). It is also suggested that CLIP’s learning process may capture how abstract concepts penetrate early vision in a top-down manner.…”
Section: Introductionmentioning
confidence: 99%