Findings of the Association for Computational Linguistics: EMNLP 2021 2021
DOI: 10.18653/v1/2021.findings-emnlp.334
|View full text |Cite
|
Sign up to set email alerts
|

A Plug-and-Play Method for Controlled Text Generation

Abstract: Large pre-trained language models have repeatedly shown their ability to produce fluent text. Yet even when starting from a prompt, generation can continue in many plausible directions. Current decoding methods with the goal of controlling generation, e.g., to ensure specific words are included, either require additional models or fine-tuning, or work poorly when the task at hand is semantically unconstrained, e.g., story generation. In this work, we present a plug-and-play decoding method for controlled langu… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
10
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 26 publications
(10 citation statements)
references
References 20 publications
0
10
0
Order By: Relevance
“…For example, PPCM (Madotto et al 2020) updates the hidden state in direction of the attribute enhancement to generate attribute-aware conversations. Pascual (Pascual et al 2021) designed a complex plug-and-play architecture to ensure that the generated content contains specific keywords. GeDi (Krause et al 2021) and its extension (Lin and Riedl 2021) can accelerate the decoding process of the PPLM, but they assume that the model is trained using large-scale labeled datasets, which are unrealizable for text infilling.…”
Section: Constrained Text Generationmentioning
confidence: 99%
“…For example, PPCM (Madotto et al 2020) updates the hidden state in direction of the attribute enhancement to generate attribute-aware conversations. Pascual (Pascual et al 2021) designed a complex plug-and-play architecture to ensure that the generated content contains specific keywords. GeDi (Krause et al 2021) and its extension (Lin and Riedl 2021) can accelerate the decoding process of the PPLM, but they assume that the model is trained using large-scale labeled datasets, which are unrealizable for text infilling.…”
Section: Constrained Text Generationmentioning
confidence: 99%
“…A representative of this group of methods is PPLM [Dathathri et al, 2020], which first trains an attribute discriminant model and then uses it to guide language model to generate the text with corresponding topic or sentiment. This group also includes the Keyword2Text method [15], which can be applied to an existing autoregressive language model without additional training. The idea of the method is to shift the output distribution of the language generation model to the semantic space of a given guide word in the word2vec or GloVe vector space.…”
Section: Previous Workmentioning
confidence: 99%
“…A representative controllable PLM is the Plug and Play Language Model [31], also called PPLM, which combines a PLM with one or more simple attribute classifiers that guide text generation without any further training of the PLM. Several studies achieved the goal of controllablility from a distributional view [89,140]. For example, Pascual et al [140] present a plug-and-play decoding method, which can be described in a single sentence: given a topic or keyword, the model add a shift to the probability distribution over the vocabulary towards semantically similar words.…”
Section: Optimization Viewmentioning
confidence: 99%
“…Several studies achieved the goal of controllablility from a distributional view [89,140]. For example, Pascual et al [140] present a plug-and-play decoding method, which can be described in a single sentence: given a topic or keyword, the model add a shift to the probability distribution over the vocabulary towards semantically similar words.…”
Section: Optimization Viewmentioning
confidence: 99%