Findings of the Association for Computational Linguistics: ACL 2022 2022
DOI: 10.18653/v1/2022.findings-acl.105
|View full text |Cite
|
Sign up to set email alerts
|

DuReadervis: A : A Chinese Dataset for Open-domain Document Visual Question Answering

Abstract: Open-domain question answering has been used in a wide range of applications, such as web search and enterprise search, which usually takes clean texts extracted from various formats of documents (e.g., web pages, PDFs, or Word documents) as the information source. However, designing different text extraction approaches is time-consuming and not scalable. In order to reduce human cost and improve the scalability of QA systems, we propose and study an Open-domain Document Visual Question Answering (Open-domain … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
3
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
3
2
2

Relationship

0
7

Authors

Journals

citations
Cited by 8 publications
(3 citation statements)
references
References 16 publications
0
3
0
Order By: Relevance
“…The authors were successful in creating a novel dataset and fundamental English methodologies. Inspired by that success, various further studies have been created and implemented in a variety of languages [6] including Chinese [7], Japanese [8], and Vietnamese [9].…”
Section: Background and Related Workmentioning
confidence: 99%
“…The authors were successful in creating a novel dataset and fundamental English methodologies. Inspired by that success, various further studies have been created and implemented in a variety of languages [6] including Chinese [7], Japanese [8], and Vietnamese [9].…”
Section: Background and Related Workmentioning
confidence: 99%
“…There are however few text-only QA datasets that contain such answer types [69,46,18]. Other datasets mainly related to our work are rather domain-specific like [112,87,56,86,73]. We give a detailed comparison of most related document VQA datasets in Table 1 highlighting the major contributions.…”
Section: Related Workmentioning
confidence: 99%
“…DocVQA (Mathew, Karatzas, and Jawahar 2021) performs question answering on industry documents. VisualMRC (Tanaka, Nishida, and Yoshida 2021), WebSRC (Chen et al 2021), WebQA (Chang et al 2022) and DuReader vis (Qi et al 2022) require comprehension on web pages. InfographicVQA (Mathew et al 2022) focuses on arithmetic reasoning over infographics.…”
Section: Multimodal Question Answeringmentioning
confidence: 99%