Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conferen 2019
DOI: 10.18653/v1/d19-1070
|View full text |Cite
|
Sign up to set email alerts
|

Effective Use of Transformer Networks for Entity Tracking

Abstract: Tracking entities in procedural language requires understanding the transformations arising from actions on entities as well as those entities' interactions. While self-attention-based pre-trained language encoders like GPT and BERT have been successfully applied across a range of natural language understanding tasks, their ability to handle the nuances of procedural texts is still untested. In this paper, we explore the use of pre-trained transformer networks for entity tracking tasks in procedural text. Firs… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
21
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
4
2

Relationship

0
6

Authors

Journals

citations
Cited by 17 publications
(22 citation statements)
references
References 16 publications
1
21
0
Order By: Relevance
“…This is "pretrained dynamics," we also consider a version without a randomly initialized dynamics model. e. (Gupta and Durrett, 2019)-style. Thiso paper proposes using Transformers to model physical state, for tasks like entity tracking in recipes.…”
Section: Pigpen-nlu Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…This is "pretrained dynamics," we also consider a version without a randomly initialized dynamics model. e. (Gupta and Durrett, 2019)-style. Thiso paper proposes using Transformers to model physical state, for tasks like entity tracking in recipes.…”
Section: Pigpen-nlu Resultsmentioning
confidence: 99%
“…PIGLeT also outperforms 'BERT style' approaches that control for the same language model architecture, but perform the physical reasoning inside the language transformer rather than as a separate model. Performance drops when the physical decoder must be learned from few paired examples (as in Gupta and Durrett (2019)); it drops even further when neither model is given access to our pretrained dynamics model, with both baselines then underperforming 'No Change.' This suggests that our approach of having a physical reasoning model outside of an LM is a good inductive bias.…”
Section: Pigpen-nlu Resultsmentioning
confidence: 99%
“…Transformer architectures trained on language modeling have been recently adapted to downstream tasks demonstrating state-of-the-art performance (Weller and Seppi, 2019;Gupta and Durrett, 2019;Maronikolakis et al, 2020). In this paper, we adapt and subsequently combine transformers with external linguistic information for complaint prediction.…”
Section: Transformer-based Modelsmentioning
confidence: 99%
“…Thus, newer models are not particularly different from the perspective of the NLI task itself. Following existing work [19,49], we include a comparison of models trained from the same BERT BAS E checkpoint. Table 4 shows the accuracy of the classification-only model and our multi-task trained models on the MNLI dataset, all having the same BERT BAS E as a starting point.…”
Section: Performance On Original Nli Taskmentioning
confidence: 99%