2019
DOI: 10.48550/arxiv.1904.03746
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Unsupervised Recurrent Neural Network Grammars

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
23
0

Year Published

2019
2019
2021
2021

Publication Types

Select...
8

Relationship

0
8

Authors

Journals

citations
Cited by 11 publications
(23 citation statements)
references
References 61 publications
0
23
0
Order By: Relevance
“…RL-trained models are finetuned from XEtrained model, optimizing CIDEr score with REIN-FORCE (Williams, 1992). The baseline used RE-INFORCE algorithm follows (Mnih and Rezende, 2016;Kim et al, 2019). This baseline performs better than self critical baseline in (Rennie et al, 2017).…”
Section: Appendices a Training Detailsmentioning
confidence: 99%
“…RL-trained models are finetuned from XEtrained model, optimizing CIDEr score with REIN-FORCE (Williams, 1992). The baseline used RE-INFORCE algorithm follows (Mnih and Rezende, 2016;Kim et al, 2019). This baseline performs better than self critical baseline in (Rennie et al, 2017).…”
Section: Appendices a Training Detailsmentioning
confidence: 99%
“…It is noteworthy that there are many other important miscellaneous works we do not mention in the previous sections. For example, numerous works have proposed to improve upon vanilla gradient-based methods [174,178,65]; linguistic rules such as negation, morphological inflection can be extracted by neural models [141,142,158]; probing tasks can used to explore linguistic properties of sentences [3,80,43,75,89,74,34]; the hidden state dynamics in recurrent nets are analysed to illuminate the learned long-range dependencies [73,96,67,179,94]; [169,166,168,101,57,167] studied the ability of neural sequence models to induce lexical, grammatical and syntactic structures; [91,90,12,136,159,24,151,85] modeled the reasoning process of the model to explain model behaviors; [157,139,28,163,219,170,180,137,106,58,162,81...…”
Section: Miscellaneousmentioning
confidence: 99%
“…Most other work focuses on constituency parsing [4], [2], [3], [21]. These methods are also bottom-up, i.e., they can extract internal data structures without needing to understand the full context of the sentence.…”
Section: Related Workmentioning
confidence: 99%