2021
DOI: 10.48550/arxiv.2108.10566
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

sigmoidF1: A Smooth F1 Score Surrogate Loss for Multilabel Classification

Abstract: Multiclass multilabel classification refers to the task of attributing multiple labels to examples via predictions. Current models formulate a reduction of that multilabel set-

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
6
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
1
1

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(7 citation statements)
references
References 38 publications
1
6
0
Order By: Relevance
“…Gamma correction helped in making the distinguishable features more evident, which was also evident in similar other studies [23]. A better understanding of the machine-learning performance can be analyzed using the F1-score, which is calculated using precision and sensitivity, especially for multiclass problems [64]. The best performing combination was the ResNet50 network using the original thermogram images with a weighted F1-score of 76.66%, followed by the MobileNetv2 network (75.74%) using the AHE-enhanced thermogram, ResNet18 using the original thermogram (75.61%), and ResNet18 and ResNet50 using the Gamma-enhanced thermogram provided scores of 74.41%, and 74.17%, respectively.…”
Section: Discussionsupporting
confidence: 53%
“…Gamma correction helped in making the distinguishable features more evident, which was also evident in similar other studies [23]. A better understanding of the machine-learning performance can be analyzed using the F1-score, which is calculated using precision and sensitivity, especially for multiclass problems [64]. The best performing combination was the ResNet50 network using the original thermogram images with a weighted F1-score of 76.66%, followed by the MobileNetv2 network (75.74%) using the AHE-enhanced thermogram, ResNet18 using the original thermogram (75.61%), and ResNet18 and ResNet50 using the Gamma-enhanced thermogram provided scores of 74.41%, and 74.17%, respectively.…”
Section: Discussionsupporting
confidence: 53%
“…Considering LSTM, in the case of the wrapping machine the only hyperparameters that vary are the type of LSTM layer, Unidirectional or Bidirectional, and the loss function. For the latter, we compare the sigmoidF 1 loss with β = 1 and η = 0 (from now on referred to as F 1 loss) [77] and the Binary Cross-Entropy (BCE) loss. The use of BiLSTM has been shown to be effective in the works [74,78], which focus on forecasting problems.…”
Section: Algorithms and Hyperparameter Tuningmentioning
confidence: 99%
“…The use of BiLSTM has been shown to be effective in the works [74,78], which focus on forecasting problems. The two compared loss functions have been proven to deliver the top performances in [77]. Each loss function serves a distinct purpose and offers advantages and disadvantages.…”
Section: Algorithms and Hyperparameter Tuningmentioning
confidence: 99%
“…Surrogate loss functions attempt to mimic certain aspects of the F β and is another related area. For example, sigmoidF1 from [1] creates smooth versions for the entries of the confusion matrix, which is used to create a differentiable loss function that imitates the F 1 . This smooth differentiability is another application of a sigmoid approximation similar to [12].…”
Section: Related Workmentioning
confidence: 99%
“…They also motivate the rationale on using statistical distributions to understand the CE curve. The van Rijsbergen's effectiveness measure is given in (1).…”
Section: Introductionmentioning
confidence: 99%