“…Research in dialogue generation has rapidly evolved from sequence-to-sequence (Sutskever et al, 2014) and Transformer models (Vaswani et al, 2017) to approaches with pre-trained models such as BERT (Devlin et al, 2019), XLNet and T5 (Raffel et al, 2020). More recently, it included techniques that use knowledge, in addition to the original posts, to improve the quality of the generated responses (Ghazvininejad et al (2018), Moghe et al (2018), Dinan et al (2019), Galley et al (2019), Lian et al (2019), Zheng and Zhou (2019), Zhao et al (2020a), Zhao et al (2020b)). 1 This approach is referred to as…”