2023
DOI: 10.1016/j.ipm.2023.103264
|View full text |Cite
|
Sign up to set email alerts
|

TSVFN: Two-Stage Visual Fusion Network for multimodal relation extraction

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
3
2
2

Relationship

0
7

Authors

Journals

citations
Cited by 25 publications
(2 citation statements)
references
References 5 publications
0
2
0
Order By: Relevance
“…Numerous efforts have been devoted to extracting relational facts from multimodal data and developing related datasets. Notably, MNRE [33,34] is the first multimodal relation extraction dataset, with a series of work demonstrating that incorporating multimodal information and efficient alignment strategy for textual and visual representations can effectively boost the performance of relation extraction in social media texts [26,27,29,32]. Wan et al [22] In future work, we will explore ways to enhance the discriminative ability of learned representations [11] and optimally fuse the complementary information of each modal to uncover the intrinsic structure of multimodal relational facts [23,24].…”
Section: Related Workmentioning
confidence: 99%
“…Numerous efforts have been devoted to extracting relational facts from multimodal data and developing related datasets. Notably, MNRE [33,34] is the first multimodal relation extraction dataset, with a series of work demonstrating that incorporating multimodal information and efficient alignment strategy for textual and visual representations can effectively boost the performance of relation extraction in social media texts [26,27,29,32]. Wan et al [22] In future work, we will explore ways to enhance the discriminative ability of learned representations [11] and optimally fuse the complementary information of each modal to uncover the intrinsic structure of multimodal relational facts [23,24].…”
Section: Related Workmentioning
confidence: 99%
“…Latent Type Classifier 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 1 0 2022b; , multimodal relation extraction (Zhao et al, 2023;, and multimodal entity linking (Zhang et al, 2023;Wang et al, 2022c). Information extraction techniques that incorporate multimodality form the foundation for constructing multimodal knowledge bases, providing ample data support for applications such as question-answering systems, information retrieval, and more.…”
Section: Known Type Classifiermentioning
confidence: 99%