“…Procedures that force LMs to be more focused on a prompt, or a specific part of it, when generating or ranking tokens can benefit algorithms that search for combinations of words through sampling. It would be interesting to use coherence boosting models in non-autoregressive text generation algorithms, such as to accelerate the mixing of MCMC methods for constrained text generation (e.g., Miao et al (2019); Zhang et al (2020a); Malkin et al (2021)). (Holtzman et al, 2021) is an unconditional probability normalization method, CC (Zhao et al, 2021) is the contextual calibration method and Channel (Min et al, 2021) uses an inverted-LM scoring approach that computes the conditional probability of the input given the label.…”