Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics 2020
DOI: 10.18653/v1/2020.acl-main.398
|View full text |Cite
|
Sign up to set email alerts
|

TaPas: Weakly Supervised Table Parsing via Pre-training

Abstract: Answering natural language questions over tables is usually seen as a semantic parsing task. To alleviate the collection cost of full logical forms, one popular approach focuses on weak supervision consisting of denotations instead of logical forms. However, training semantic parsers from weak supervision poses difficulties, and in addition, the generated logical forms are only used as an intermediate step prior to retrieving the denotation. In this paper, we present TAPAS, an approach to question answering ov… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
456
2

Year Published

2020
2020
2023
2023

Publication Types

Select...
4
1
1

Relationship

1
5

Authors

Journals

citations
Cited by 352 publications
(458 citation statements)
references
References 31 publications
0
456
2
Order By: Relevance
“…On SQA (Iyyer et al, 2017) a model pre-trained on the synthetic entailment data outperforms one pre-trained on the MASK-LM task alone (Table 3). Our best BERT Base model outpeforms the BERT-Large model of Herzig et al (2020) and a BERT-Large model trained on our data improves the previous state-of-the-art by 4 points on average question and sequence accuracy. See dev results and error bars in Appendix E. Efficiency As discussed in Section 3.3 and Appendix A.4, we can increase the model efficiency by reducing the input length.…”
Section: Zero-shot Accuracy and Low Resource Regimesmentioning
confidence: 75%
See 4 more Smart Citations
“…On SQA (Iyyer et al, 2017) a model pre-trained on the synthetic entailment data outperforms one pre-trained on the MASK-LM task alone (Table 3). Our best BERT Base model outpeforms the BERT-Large model of Herzig et al (2020) and a BERT-Large model trained on our data improves the previous state-of-the-art by 4 points on average question and sequence accuracy. See dev results and error bars in Appendix E. Efficiency As discussed in Section 3.3 and Appendix A.4, we can increase the model efficiency by reducing the input length.…”
Section: Zero-shot Accuracy and Low Resource Regimesmentioning
confidence: 75%
“…We use a model architecture derived from BERT and add additional embeddings to encode the table structure, following the approach of Herzig et al (2020) to encode the input.…”
Section: Modelmentioning
confidence: 99%
See 3 more Smart Citations