2020 IEEE 14th International Conference on Semantic Computing (ICSC) 2020
DOI: 10.1109/icsc.2020.00018
|View full text |Cite
|
Sign up to set email alerts
|

Early Forecasting of Text Classification Accuracy and F-Measure with Active Learning

Abstract: When creating text classification systems, one of the major bottlenecks is the annotation of training data. Active learning has been proposed to address this bottleneck using stopping methods to minimize the cost of data annotation. An important capability for improving the utility of stopping methods is to effectively forecast the performance of the text classification models. Forecasting can be done through the use of logarithmic models regressed on some portion of the data as learning is progressing. A crit… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2020
2020
2020
2020

Publication Types

Select...
2

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(1 citation statement)
references
References 28 publications
0
1
0
Order By: Relevance
“…Active learning, which iteratively selects the most informative samples for labelers to annotate, is an effective method to reduce annotation cost [5]- [7]. It has been widely used in many Natural Language Processing (NLP) tasks, such as text classification [8] and event recognition [9]. In conventional active learning, there is only one labeler and the algorithm queries the labels of the selected instances from the labeler, which always returns the ground truth of queried labels [10].…”
Section: Introductionmentioning
confidence: 99%
“…Active learning, which iteratively selects the most informative samples for labelers to annotate, is an effective method to reduce annotation cost [5]- [7]. It has been widely used in many Natural Language Processing (NLP) tasks, such as text classification [8] and event recognition [9]. In conventional active learning, there is only one labeler and the algorithm queries the labels of the selected instances from the labeler, which always returns the ground truth of queried labels [10].…”
Section: Introductionmentioning
confidence: 99%