2014
DOI: 10.4018/ijmcmc.2014100102
|View full text |Cite
|
Sign up to set email alerts
|

A Strategy on Selecting Performance Metrics for Classifier Evaluation

Abstract: The evaluation of classifiers' performances plays a critical role in construction and selection of classification model. Although many performance metrics have been proposed in machine learning community, no general guidelines are available among practitioners regarding which metric to be selected for evaluating a classifier's performance. In this paper, we attempt to provide practitioners with a strategy on selecting performance metrics for classifier evaluation. Firstly, the authors investigate seven widely … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
52
0

Year Published

2016
2016
2023
2023

Publication Types

Select...
4
4
2

Relationship

4
6

Authors

Journals

citations
Cited by 140 publications
(52 citation statements)
references
References 26 publications
0
52
0
Order By: Relevance
“…In order to ensure the effectiveness of our experiment, we divide the dataset into training, development and test in the ratio of 8:1:1. In the following experiments, we use common performance measures such as Precision, Recall, and F 1 -score [19] to evaluate NER, RC and joint models.…”
Section: Experimental Studies a Dataset And Evaluation Metricsmentioning
confidence: 99%
“…In order to ensure the effectiveness of our experiment, we divide the dataset into training, development and test in the ratio of 8:1:1. In the following experiments, we use common performance measures such as Precision, Recall, and F 1 -score [19] to evaluate NER, RC and joint models.…”
Section: Experimental Studies a Dataset And Evaluation Metricsmentioning
confidence: 99%
“…Clinical text structuring is an important problem which is highly related to practical applications. Considerable efforts arXiv:1908.06606v2 [cs.CL] 22 Oct 2019 have been made on CTS task. These studies can be roughly divided into three categories, namely rule and dictionary based methods, task-specific end-to-end methods and pipeline methods.…”
Section: Related Work a Clinical Text Structuringmentioning
confidence: 99%
“…We further split these sentences into clauses by commas. Detailed statistics of different types of entities are listed in Table II. In the following experiments, widely-used performance measures such as precision, recall, and F 1 -score [40], [41] are used to evaluate the methods.…”
Section: A Dataset and Evaluation Metricsmentioning
confidence: 99%