Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP) 2020
DOI: 10.18653/v1/2020.emnlp-main.577
|View full text |Cite
|
Sign up to set email alerts
|

An Unsupervised Joint System for Text Generation from Knowledge Graphs and Semantic Parsing

Abstract: Knowledge graphs (KGs) can vary greatly from one domain to another. Therefore supervised approaches to both graph-to-text generation and text-to-graph knowledge extraction (semantic parsing) will always suffer from a shortage of domain-specific parallel graphtext data; at the same time, adapting a model trained on a different domain is often impossible due to little or no overlap in entities and relations. This situation calls for an approach that (1) does not need large amounts of annotated data and thus (2) … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
35
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
1

Relationship

0
6

Authors

Journals

citations
Cited by 29 publications
(35 citation statements)
references
References 42 publications
0
35
0
Order By: Relevance
“…Dataset Overview For our two versions of datasets, GenWiki FINE and GenWiki FULL , we summarize the overall statistics of our GenWiki dataset in Table 3. We compare it with WebNLG, which also meets all three basic requirements, and has been used for previous unsupervised models (Guo et al, 2020;Schmitt et al, 2020 We can see from Table 3 that our dataset has significantly more data than the human-annotated WebNLG, and can be a better dataset for unsupervised learning. Our GenWiki FINE contains 757K ex-amples with about 20M tokens, and GenWiki FULL contains 1.3M samples with about 30M tokens.…”
Section: Discussionmentioning
confidence: 99%
See 4 more Smart Citations
“…Dataset Overview For our two versions of datasets, GenWiki FINE and GenWiki FULL , we summarize the overall statistics of our GenWiki dataset in Table 3. We compare it with WebNLG, which also meets all three basic requirements, and has been used for previous unsupervised models (Guo et al, 2020;Schmitt et al, 2020 We can see from Table 3 that our dataset has significantly more data than the human-annotated WebNLG, and can be a better dataset for unsupervised learning. Our GenWiki FINE contains 757K ex-amples with about 20M tokens, and GenWiki FULL contains 1.3M samples with about 30M tokens.…”
Section: Discussionmentioning
confidence: 99%
“…Some more specific constraints require, for example, the text corpus to have entity annotations. The reason is that recent unsupervised learning models (Guo et al, 2020;Schmitt et al, 2020) use cycle training of two tasks: graph-to-text, and text-to-graph, which is simplified to relation extraction given entities. This simplification requires the unsupervised text corpus to have entity annotations in text.…”
Section: Desideratamentioning
confidence: 99%
See 3 more Smart Citations