2020
DOI: 10.1109/access.2020.3023421
|View full text |Cite
|
Sign up to set email alerts
|

Can Machines Tell Stories? A Comparative Study of Deep Neural Language Models and Metrics

Abstract: Massive textual content has enabled rapid advances in natural language modeling. The use of pre-trained deep neural language models has significantly improved natural language understanding tasks. However, the extent to which these systems can be applied to content generation is unclear. While a few informal studies have claimed that these models can generate 'high quality' readable content, there is no prior study on analyzing the generated content from these models based on sampling and fine-tuning hyperpara… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
9
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
4
1

Relationship

0
9

Authors

Journals

citations
Cited by 19 publications
(9 citation statements)
references
References 32 publications
0
9
0
Order By: Relevance
“…The metric score of the content generated by the models was compared with the metric score of human written content; no model was close to human written content. However, the fine-tuned GPT-2 models and the transformer-extra long (XL) models were the ones that had lowest deviation in the metrics whereas the XL Net models had a high value of deviation [10]. Data imbalance is one of the major problems in classification, lack of a dataset belonging to any of the class labels could result in data imbalance and further lead to poor classification performance results.…”
Section: Literature Surveymentioning
confidence: 99%
“…The metric score of the content generated by the models was compared with the metric score of human written content; no model was close to human written content. However, the fine-tuned GPT-2 models and the transformer-extra long (XL) models were the ones that had lowest deviation in the metrics whereas the XL Net models had a high value of deviation [10]. Data imbalance is one of the major problems in classification, lack of a dataset belonging to any of the class labels could result in data imbalance and further lead to poor classification performance results.…”
Section: Literature Surveymentioning
confidence: 99%
“…Perplexity of LM. Perplexity is commonly used to evaluate probabilistic models, particularly LMs [7]. In the context of an LM, perplexity measures the average likelihood of its predictions for the next token.…”
Section: Dt Rmentioning
confidence: 99%
“…As all these models are pretrained and publicly available, they represent a significant threat in the field of misleading reviews, as anyone can retrain the model on a specific field and use them for malicious purposes. GPT2 has been used extensively for text generation (Das and Verma, 2020) and review generation (Salminen et al, 2022). BERT (Bidirectional Encoding Representations of Transformers) is another transformer-based model that has been frequently utilized for language interpretation tasks, although some recent works have employed this model for text generation (Devlin et al, 2019).…”
Section: Review Of Text Generatorsmentioning
confidence: 99%