2019
DOI: 10.1016/j.csl.2018.08.006
|View full text |Cite
|
Sign up to set email alerts
|

Adversarial training and decoding strategies for end-to-end neural conversation models

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
4
0
3

Year Published

2020
2020
2023
2023

Publication Types

Select...
5
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 12 publications
(7 citation statements)
references
References 10 publications
0
4
0
3
Order By: Relevance
“…Automat ic based Metric F1-Score [89], [114], [106], [80], [62], [100] Precision [105], [106], [80], [62] Recall [105], [106], [80], [62] Accuracy [92], [114], [126], [105], [124], [80], [ 93], [62], [100], [76], [56], [74] PPL [89], [102], [94], [92], [98], [124], [12 6], [39], [101], [85] BLEU [79], [83], [94], [81], [92], [95], [84], [ 126], [127], [86], [101], [49], [40],[4 0],…”
Section: Categori Zation Metrics Articlesmentioning
confidence: 99%
“…Automat ic based Metric F1-Score [89], [114], [106], [80], [62], [100] Precision [105], [106], [80], [62] Recall [105], [106], [80], [62] Accuracy [92], [114], [126], [105], [124], [80], [ 93], [62], [100], [76], [56], [74] PPL [89], [102], [94], [92], [98], [124], [12 6], [39], [101], [85] BLEU [79], [83], [94], [81], [92], [95], [84], [ 126], [127], [86], [101], [49], [40],[4 0],…”
Section: Categori Zation Metrics Articlesmentioning
confidence: 99%
“…Prior to the training of the dialogue system, we pretrain the word vector matrix in the same corpus that will be used later. Following the work of [14] and [9], we also pretrain the generator using the MLE criteria, and the discriminator with the responses generated by the pretrained generator and with responses from the corpus. In order to stabilize the rest of the training process and to avoid the catastrophic forgetting phenomenon of the discriminator, each time we sample a response of the generator to a given input, we add it to a corpus of generator's turns C D .…”
Section: Training Proceduresmentioning
confidence: 99%
“…We finally repeat this process of training the generator, adding samples to C D and training the discriminator, but this time training the corpus with the MLE criteria. This approach is also taken in [14] and [9], and it aims at stabilizing the training process. In order to further stabilize it, we reduce the learning rate of the training optimizer throughout the global iterations.…”
Section: Training Proceduresmentioning
confidence: 99%
See 1 more Smart Citation
“…Xehetasunetan sartu gabe, sare diskriminatzailearen irteera ez da diferentziagarria sare sortzailearen parametroekiko, sortzaileak sortutako hitzak diskretuak dira eta [11]. Errefortzu bidezko ikasketa erabili daiteke gradienteetan oinarritutako metodoen ordez [12,13], baina horrek entrenamenduaren konbergentzia zaildu dezake [14]. Beste aukera bat straight-through Gumbel-softmax [15,16] zenbateslearen bidez gradientearen hurbilketa bat egitea da, [17] eta [18] laneetan erakusten duten moduan.…”
Section: Arloko Egoera Eta Ikerketaren Helburuakunclassified