2021
DOI: 10.1002/hbm.25603
|View full text |Cite
|
Sign up to set email alerts
|

A dual‐channel language decoding from brain activity with progressive transfer training

Abstract: When we view a scene, the visual cortex extracts and processes visual information in the scene through various kinds of neural activities. Previous studies have decoded the neural activity into single/multiple semantic category tags which can caption the scene to some extent. However, these tags are isolated words with no grammatical structure, insufficiently conveying what the scene contains. It is well-known that textual language (sentences/phrases) is superior to single word in disclosing the meaning of ima… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

1
5
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
2

Relationship

2
4

Authors

Journals

citations
Cited by 7 publications
(6 citation statements)
references
References 40 publications
1
5
0
Order By: Relevance
“…Third, our multitask visual-language decoding model succeeded in accurately capturing key objects and actions within images, organizing them into fluent and comprehensible textual descriptions. This echoes well with our prior work which directly utilized a dual-channel language model to generate precise language information from fMRI neural activity 20 . Furthermore, our findings indirectly corroborated that higher brain areas, such as FFA and VWFA, are primarily responsible for handling high-level semantic information, while lower-level brain regions like V1 and V2 handle fundamental visual information 41–43 .…”
Section: Discussionsupporting
confidence: 84%
See 2 more Smart Citations
“…Third, our multitask visual-language decoding model succeeded in accurately capturing key objects and actions within images, organizing them into fluent and comprehensible textual descriptions. This echoes well with our prior work which directly utilized a dual-channel language model to generate precise language information from fMRI neural activity 20 . Furthermore, our findings indirectly corroborated that higher brain areas, such as FFA and VWFA, are primarily responsible for handling high-level semantic information, while lower-level brain regions like V1 and V2 handle fundamental visual information 41–43 .…”
Section: Discussionsupporting
confidence: 84%
“…Further, to investigate preferences within the visual cortex for language decoding, we assessed ten distinct regions (including V1, V2, V3, EBA, FBA, FFA, OFA, OPA, PPA, and VWFA) along with three combined areas (LVC, HVC, and VC) for their decoding accuracy using Word2vec. We chose Word2vec due to its demonstrated effectiveness in assessing language relevance and measuring semantic similarity 20 .…”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…To do so, we introduce a quantitative framework investigating the two dimensions of dialogues’ emotional content and syntactic/semantic structure for predicting interpersonal neu-ral synchrony during naturalistic social exchanges. Our study, thus, bridges innovative cognitive neuroscience data [40] with affective computing and cognitive data science frameworks [30], integrating mind and brain data [41, 42]. In particular, we lever-age the recent modeling framework of Textual Forma Mentis Networks (TFMNs) for representing conceptual associations encoded in text and retrieved via artificial intelli-gence (AI) and psychologically validated data [32, 43].…”
Section: Introductionmentioning
confidence: 99%
“…Convolutional neural networks (CNNs) have been widely used in image recognition (Russakovsky et al, 2015 ; He et al, 2016 ), object detection (Lin et al, 2014 ), speech recognition (Yu et al, 2016 ), visual coding and decoding (Huang et al, 2021a , b ). However, traditional CNNs can only handle data in the Euclidean space.…”
Section: Introductionmentioning
confidence: 99%