2020
DOI: 10.3938/jkps.76.368
|View full text |Cite
|
Sign up to set email alerts
|

Multi-Label Classification of Historical Documents by Using Hierarchical Attention Networks

Abstract: The quantitative analysis of digitized historical documents has begun in earnest in recent years. Text classification is of particular importance for quantitative historical analysis because it helps to search literature efficiently and to determine the important subjects of a particular age. While numerous historians have joined together to classify large-scale historical documents, consistent classification among individual researchers has not been achieved. In this study, we present a classification method … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
3
1

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(3 citation statements)
references
References 9 publications
0
3
0
Order By: Relevance
“…However, by stemming the vocabulary is significantly reduced and therefore the training time is also reduced. Kim [67] performed a multi-label classification on a Korean translated dataset and achieved accuracy up to 71%. Huang et al [68] used four datasets, Amazon Mobile Phone reviews, Amazon fine food reviews, and Yelp reviews 1 and 2.…”
Section: Deep Learning-based Approachesmentioning
confidence: 99%
“…However, by stemming the vocabulary is significantly reduced and therefore the training time is also reduced. Kim [67] performed a multi-label classification on a Korean translated dataset and achieved accuracy up to 71%. Huang et al [68] used four datasets, Amazon Mobile Phone reviews, Amazon fine food reviews, and Yelp reviews 1 and 2.…”
Section: Deep Learning-based Approachesmentioning
confidence: 99%
“…Similarly, the attention structure is fed to allow the system to emphasize the LSTM or GRU outcomes accompanying with the arguments and lines that are most revealing of a specific class. We then established both models, i.e., LSTMs and GRUs through the hypermeter optimization process [40].…”
Section: Classificationmentioning
confidence: 99%
“…HANs are currently one of the most popular neural network algorithms being adopted in Computer Science. In 2020, the year of publication of this research, we have seen their applications in many different fields, such as Finance, to predict stock market values - Huang et al (2020); Linguistics, in the classification of historical documents - Kim et al (2020); Technology, in mobile app recommendations - Liang et al (2020); and Medicine, with detection models for Atrial Fibrillation -Mousavi, Afghah and Acharya (2020). We believe this flexibility shows the strength of the Attention model to handle different texts across many different vocabularies.…”
Section: Hierarchical Attention Network -Hanmentioning
confidence: 99%