2015
DOI: 10.1007/978-3-319-19581-0_1
|View full text |Cite
|
Sign up to set email alerts
|

Improving Supervised Classification Using Information Extraction

Abstract: Abstract. We explore supervised learning for multi-class, multi-label text classification, focusing on real-world settings, where the distribution of labels changes dynamically over time. We use the PULS Information Extraction system to collect information about the distribution of class labels over named entities found in text. We then combine a knowledge-based rote classifier with statistical classifiers to obtain better performance than either classification method alone. The resulting classifier yields a s… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

1
3
0

Year Published

2016
2016
2021
2021

Publication Types

Select...
2
1

Relationship

2
1

Authors

Journals

citations
Cited by 3 publications
(4 citation statements)
references
References 23 publications
(24 reference statements)
1
3
0
Order By: Relevance
“…Compared to the reported state-of-the-art results on Sector Classification (Table 2), our best model yields a 10% gain in µ-F1, (Cisse et al, 2013), and a 6% gain in M-F1 (Du et al, 2015). The best µ-F1 and M-F1 results are obtained by the same model.…”
Section: Algorithm (Prior)supporting
confidence: 51%
See 2 more Smart Citations
“…Compared to the reported state-of-the-art results on Sector Classification (Table 2), our best model yields a 10% gain in µ-F1, (Cisse et al, 2013), and a 6% gain in M-F1 (Du et al, 2015). The best µ-F1 and M-F1 results are obtained by the same model.…”
Section: Algorithm (Prior)supporting
confidence: 51%
“…To the best of our knowledge, our previous work (Du et al, 2015) was the only study of the utility of NEs for RCV1 classification. We demonstrated that using a combination of keyword-based and NE-based classifiers works better than either classifier alone.…”
Section: Data and Prior Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Additionally, because subword tokenization is not used, their motivation differs from ours that investigates the effectiveness of NEs on subword-based neural networks. Du et al (2015) investigated the use of NEs in non-neural network models on document classification. While they reported use of NEs improve the accuracy of document classification, the contribution to subword-based neural network models was not investigated.…”
Section: Related Workmentioning
confidence: 99%