Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics 2020
DOI: 10.18653/v1/2020.acl-main.117
|View full text |Cite
|
Sign up to set email alerts
|

The TechQA Dataset

Abstract: We introduce TECHQA, a domain-adaptation question answering dataset for the technical support domain. The TECHQA corpus highlights two real-world issues from the automated customer support domain. First, it contains actual questions posed by users on a technical forum, rather than questions generated specifically for a competition or a task. Second, it has a real-world size -600 training, 310 dev, and 490 evaluation question/answer pairs -thus reflecting the cost of creating large labeled datasets with actual … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
25
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
2
1

Relationship

1
6

Authors

Journals

citations
Cited by 21 publications
(25 citation statements)
references
References 9 publications
0
25
0
Order By: Relevance
“…In this sense the system trained on the SQuAD 2.0 dataset, which is not built with prospective users, achieves highest automatic metrics but falls behind when attending to user preference. There have been recent efforts for building QA datasets closer to real-world use cases (Choi et al, 2018;Reddy et al, 2019;Campos et al, 2020;Castelli et al, 2020). Among all of them, we show that the system trained also on the QuAC dataset is preferred by the users.…”
Section: Introductionmentioning
confidence: 68%
“…In this sense the system trained on the SQuAD 2.0 dataset, which is not built with prospective users, achieves highest automatic metrics but falls behind when attending to user preference. There have been recent efforts for building QA datasets closer to real-world use cases (Choi et al, 2018;Reddy et al, 2019;Campos et al, 2020;Castelli et al, 2020). Among all of them, we show that the system trained also on the QuAC dataset is preferred by the users.…”
Section: Introductionmentioning
confidence: 68%
“…Domain Adaptation: Results on TechQA: Since VAULT has shown to be effective on NQ, we evaluate it on a new domain, TechQA. We compare it against a RoBERTa base model trained with the same hyper-parameters as (Castelli et al, 2020)except we use 11 epochs instead of 20. We chose base instead of large (as is used for the TechQA Results on TechQA are reported in Table 2.…”
Section: Results On Nqmentioning
confidence: 99%
“…Datasets: We experiment with two challenging "natural" MRC datasets: NQ (Kwiatkowski et al, 2019) and TechQA (Castelli et al, 2020). We provide a brief summary of the datasets and direct interested readers to the corresponding papers.…”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…In particular, after transferring the approximate RC knowledge, the BERT-based MRC model's performance is improved by 3.46%, while the BERT-based document retriever's ( [2]) performance is improved by 4.33% (both in absolute F1 score) on TechQA, a realworld low-resource MRC dataset [2]. Further, our best model also outperforms the previous best document retrieval scores [2] by 15.12% in document retrieval accuracy (DRA) on TechQA.…”
Section: Introductionmentioning
confidence: 99%