Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing 2021
DOI: 10.18653/v1/2021.emnlp-main.50
|View full text |Cite
|
Sign up to set email alerts
|

Artificial Text Detection via Examining the Topology of Attention Maps

Abstract: The impressive capabilities of recent generative models to create texts that are challenging to distinguish from the human-written ones can be misused for generating fake news, product reviews, and even abusive content. Despite the prominent performance of existing methods for artificial text detection, they still lack interpretability and robustness towards unseen models. To this end, we propose three novel types of interpretable topological features for this task based on Topological Data Analysis (TDA) whic… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
12
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
4
1

Relationship

0
10

Authors

Journals

citations
Cited by 14 publications
(12 citation statements)
references
References 23 publications
0
12
0
Order By: Relevance
“…[18][19][20][21] TDA has been used in NLP for movie genre detection, 22 textual entailment, 23 document summarization, 24 and analysis of sentence embeddings. 25 The topology of the attention layers has been leveraged for text classification, 26,27 acceptability judgments, 28 and robustness against adversarial attacks. 29…”
Section: Related Workmentioning
confidence: 99%
“…[18][19][20][21] TDA has been used in NLP for movie genre detection, 22 textual entailment, 23 document summarization, 24 and analysis of sentence embeddings. 25 The topology of the attention layers has been leveraged for text classification, 26,27 acceptability judgments, 28 and robustness against adversarial attacks. 29…”
Section: Related Workmentioning
confidence: 99%
“…Given an input text, we extract output attention matrices from Transformer LMs and follow Kushnareva et al, 2021 to compute three types of persistent features over them.…”
Section: Extracted Featuresmentioning
confidence: 99%
“…Their best performing model utilizes persistence features derived from time-delay embeddings of term frequency data. Kushnareva et al (2021) compute persistent homology of a filtered graph constructed from the attention maps of a pretrained language model and harness the features for an artificial text detection task.…”
Section: Related Workmentioning
confidence: 99%