Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers) 2022
DOI: 10.18653/v1/2022.acl-long.45
|View full text |Cite
|
Sign up to set email alerts
|

Quality Controlled Paraphrase Generation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
12
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
3
1
1

Relationship

0
8

Authors

Journals

citations
Cited by 14 publications
(12 citation statements)
references
References 0 publications
0
12
0
Order By: Relevance
“…(Liu and Soh, 2022a). This approach is similar to Bandel et al (2022); 's proposed metrics for paraphrase quality (e.g., semantic similarity, expression diversity).…”
Section: Related Workmentioning
confidence: 98%
See 1 more Smart Citation
“…(Liu and Soh, 2022a). This approach is similar to Bandel et al (2022); 's proposed metrics for paraphrase quality (e.g., semantic similarity, expression diversity).…”
Section: Related Workmentioning
confidence: 98%
“…Recent work requires texts to satisfy certain stylistic, semantic, or structural requirements, such as using formal language or expressing thoughts using a particular template (Iyyer et al, 2018;Shen et al, 2017). In paraphrase generation, methods require texts to meet certain quality criteria, such as semantic preservation and lexical diversity (Bandel et al, 2022;Yang et al, 2022) or require syntactic criteria, such as word ordering Goyal and Durrett, 2020;. The development of the Multi-Topic Paraphrase in Twitter (MultiPIT) corpus addresses quality issues in existing paraphrase datasets and facilitates the acquisition and generation of high-quality paraphrases (Dou et al, 2022).…”
Section: Related Workmentioning
confidence: 99%
“…Algorithm 2 summarises the candidate generation procedure, which takes any paraphrase model, a list of model-specific generation parameters and, optionally, a list of filters as an input (line 1). These parameters are temperature and number of beams for Pegasus, or a grid of lexical, semantic and syntactic distances for the Quality Controlled Paraphrase Generation (QCPG) (Bandel et al, 2022) model presented in Appendix E. The model generates one or more paraphrases, which are filtered before returning (lines 3-8). We describe the filtering process next.…”
Section: B1 Candidate Generationmentioning
confidence: 99%
“…In Table 7 we show that simply re-ranking the Pegasus output space with BLEURT improves SGD performance comparably with backtranslation (rows 2&4) and the robustness and generalisation improvements are maintained. Bandel et al (2022) exploit high quality examples in paraphrase corpora by conditioning the model with a string quality parameters string outlining target semantic, syntactic and lexical distances of the generated paraphrase during finetuning. At inference one must specify these parameters to obtain diverse yet high quality paraphrases.…”
Section: Filter Namementioning
confidence: 99%
“…4) Since our focus was on evaluating machine translation, we naturally chose translation for augmenting the data. However, other data augmentation techniques could seamlessly integrate into DATScore, such as using a text paraphrasing model (Bandel et al, 2022).…”
Section: Limitationsmentioning
confidence: 99%