2020
DOI: 10.1007/978-3-030-60450-9_12
|View full text |Cite
|
Sign up to set email alerts
|

Multimodal Aspect Extraction with Region-Aware Alignment Network

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
8
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
3
2

Relationship

1
8

Authors

Journals

citations
Cited by 21 publications
(8 citation statements)
references
References 17 publications
0
8
0
Order By: Relevance
“…Since contents in different modalities are often closely related, exploiting such multimodal information can help better analyze users' sentiments towards different aspects. Recent studies on multimodal ABSA mainly concentrate on simple ABSA tasks such as multimodal ATE [182,183] and multimodal ASC [184,185,186,187]. To align the information from different modalities, the text and image are often first encoded to feature representations, then some interaction networks are designed to fuse the information for making the final prediction.…”
Section: Multimodal Absamentioning
confidence: 99%
“…Since contents in different modalities are often closely related, exploiting such multimodal information can help better analyze users' sentiments towards different aspects. Recent studies on multimodal ABSA mainly concentrate on simple ABSA tasks such as multimodal ATE [182,183] and multimodal ASC [184,185,186,187]. To align the information from different modalities, the text and image are often first encoded to feature representations, then some interaction networks are designed to fuse the information for making the final prediction.…”
Section: Multimodal Absamentioning
confidence: 99%
“…Approaches for Multimodal Aspect Term Extraction (MATE). 1) RAN (Wu et al, 2020a), which aligns text with object regions by a coattention network. 2) UMT , which uses Cross-Modal Transformer to fuse text and image representations for Multimodal Named Entity Recognition (MNER).…”
Section: Compared Systemsmentioning
confidence: 99%
“…The first group are the most related approaches to multi-modal aspect terms extraction. 1) RAN (Wu et al, 2020a); a co-attention approach for aspect terms extraction in a multi-modal scenario. 2) UMT (Yu et al, 2020b); 3) OSCGA (Wu et al, 2020b), an NER approach in a multi-modal scenario based on object features with BIO tagging.…”
Section: Baselinesmentioning
confidence: 99%
“…Previous studies normally cast MALSA in social media as two independent sub-tasks: Multimodal Aspect Terms Extraction (MATE) and Multimodal Aspect Sentiment Classification (MASC). First, MATE aims to detect a set of all potential * Corresponding Author aspect terms from a free text with its accompanying image (Wu et al, 2020a). Second, MASC aims to classify the sentiment polarity of a multi-modal post towards a given aspect in textual modality (Yu and Jiang, 2019).…”
Section: Introductionmentioning
confidence: 99%