2023
DOI: 10.48550/arxiv.2303.15621
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

ChatGPT as a Factual Inconsistency Evaluator for Abstractive Text Summarization

Abstract: The performance of text summarization has been greatly boosted by pre-trained language models. A main concern of existing methods is that most generated summaries are not factually inconsistent with their source documents.To alleviate the problem, many efforts have focused on developing effective factuality evaluation metrics based on natural language inference, question answering, and syntactic dependency et al. However, these approaches are limited by either their high computational complexity or dependence … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
27
0
1

Year Published

2023
2023
2024
2024

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 21 publications
(28 citation statements)
references
References 22 publications
0
27
0
1
Order By: Relevance
“…In existing LLM-based text annotation methods [2,33,37,48], several prompt-tuning techniques have been utilized to enhance the performance of LLMs. Alizadeh et al [2] investigate the performance difference between LLMs and crowd-workers in text annotation tasks.…”
Section: Prompt-tuning For Text Annotationmentioning
confidence: 99%
“…In existing LLM-based text annotation methods [2,33,37,48], several prompt-tuning techniques have been utilized to enhance the performance of LLMs. Alizadeh et al [2] investigate the performance difference between LLMs and crowd-workers in text annotation tasks.…”
Section: Prompt-tuning For Text Annotationmentioning
confidence: 99%
“…Yang et al [34] explored the limitations of ChatGPT in query-based summarization. Luo et al [35] studied the practicality of ChatGPT as a fact-inconsistency evaluator for abstract text summarization. Recently, COT has significantly improved the inferential performance and interpretability of LLMs by decomposing multi-step problems into intermediate steps [20,36,37].…”
Section: Large Langauge Modelsmentioning
confidence: 99%
“…The study sought to provide a valuable contribution to the future progress of chatbot and language model development. Zheheng Luo et al examined the problems regarding factual inconsistencies [4] in generated summaries by utilizing large language models (LLMs). It specifically investigated the evaluation capabilities of ChatGPT under a zeroshot environment.…”
Section: Review On the Limitations Of Chatgptmentioning
confidence: 99%