2021
DOI: 10.1007/s10994-021-06023-5
|View full text |Cite
|
Sign up to set email alerts
|

Density-based weighting for imbalanced regression

Abstract: In many real world settings, imbalanced data impedes model performance of learning algorithms, like neural networks, mostly for rare cases. This is especially problematic for tasks focusing on these rare occurrences. For example, when estimating precipitation, extreme rainfall events are scarce but important considering their potential consequences. While there are numerous well studied solutions for classification settings, most of them cannot be applied to regression easily. Of the few solutions for regress… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
35
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 87 publications
(35 citation statements)
references
References 21 publications
0
35
0
Order By: Relevance
“…The dataset was resampled by using SMOGN as a data preprocessing step before the training of the regression models because the data distribution was not balanced. The SMOGN algorithm, which mixes oversampling with Gaussian noise [57], is based on the SMOTER method. The density of the outcome of the dataset after oversampling with the SMOGN approach is shown in Figure 7 together with the initial data density.…”
Section: Regression Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…The dataset was resampled by using SMOGN as a data preprocessing step before the training of the regression models because the data distribution was not balanced. The SMOGN algorithm, which mixes oversampling with Gaussian noise [57], is based on the SMOTER method. The density of the outcome of the dataset after oversampling with the SMOGN approach is shown in Figure 7 together with the initial data density.…”
Section: Regression Resultsmentioning
confidence: 99%
“…The SMOTER and the introduction of Gaussian Noise are two oversampling techniques combined in the SMOGN approach. Based on the distance to the k-nearest neighbors, SMOGN iterates through all unusual samples and selects between SMOTER's interpolation-based oversampling and Gaussian noise-based oversampling [57]. The data imbalance effect was eliminated by using the SMOTE and SMOGN techniques for classification and regression models, respectively to resample the dataset into the classes and change the relative frequency of the other labels.…”
Section: Resamplingmentioning
confidence: 99%
“…The data augmentation and class weight approaches are standards in deep learning. Data augmentation is an approach to increasing the diversity of training data [55], and class weight is used to determine the weight of each categorical variable when the dataset is unbalanced [56]. These approaches might be employed to ensure the diversity and completeness of selected projects in the imbalanced projects.…”
Section: Discussionmentioning
confidence: 99%
“…Many approaches revolve around modifications of SMOTE such as adapted to regression SMOTER [38], augmented with Gaussian Noise SMOGN [39], or [40] work extending for prediction of extremely rare values. [41] proposed DenseWeight, a method based on Kernel Density Estimation for better assessment of the relevance function for sample reweighing. [42] proposed a distribution smoothing over label (LDS) and feature space (FDS) for imbalanced regression.…”
Section: Related Workmentioning
confidence: 99%