Proceedings of the 3rd Workshop on Natural Language Processing for Conversational AI 2021
DOI: 10.18653/v1/2021.nlp4convai-1.20
|View full text |Cite
|
Sign up to set email alerts
|

Investigating Pretrained Language Models for Graph-to-Text Generation

Abstract: Graph-to-text generation aims to generate fluent texts from graph-based data. In this paper, we investigate two recent pretrained language models (PLMs) and analyze the impact of different task-adaptive pretraining strategies for PLMs in graph-to-text generation. We present a study across three graph domains: meaning representations, Wikipedia knowledge graphs (KGs) and scientific KGs. We show that approaches based on PLMs BART and T5 achieve new state-of-the-art results and that task-adaptive pretraining stra… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
34
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
1

Relationship

0
6

Authors

Journals

citations
Cited by 47 publications
(34 citation statements)
references
References 44 publications
(99 reference statements)
0
34
0
Order By: Relevance
“…Working with single-claim data, we do not need to maintain the graph's structure. Large pre-trained language models have achieved state of the art results when fine-tuned and evaluated on WebNLG [33,16,15], mainly the T5 [30]; they can disregard most structure and can be applied to one or many claims at a time. Hence, we utilise the T5 (base version) as our verbalisation model, following training and evaluation methods from these works.…”
Section: Approach Training and Validationmentioning
confidence: 99%
See 4 more Smart Citations
“…Working with single-claim data, we do not need to maintain the graph's structure. Large pre-trained language models have achieved state of the art results when fine-tuned and evaluated on WebNLG [33,16,15], mainly the T5 [30]; they can disregard most structure and can be applied to one or many claims at a time. Hence, we utilise the T5 (base version) as our verbalisation model, following training and evaluation methods from these works.…”
Section: Approach Training and Validationmentioning
confidence: 99%
“…The SEEN partition is used for both training and validation/testing, while the UNSEEN partition is kept for testing only. We follow the training setup from Ribeiro et al [33] by specifying a new prefix "translate from Graph to Text" and adding three new tokens ( H , R , and T ) that precede the claim's subject, predicate, and object, respectively.…”
Section: Approach Training and Validationmentioning
confidence: 99%
See 3 more Smart Citations