2021
DOI: 10.1016/j.ijar.2021.04.002
|View full text |Cite
|
Sign up to set email alerts
|

Instance weighting through data imprecisiation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
11
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
3

Relationship

2
5

Authors

Journals

citations
Cited by 8 publications
(12 citation statements)
references
References 17 publications
1
11
0
Order By: Relevance
“…However, from a data modeling perspective, it is important to recognize that the imprecisiation of the target information pays off with higher validity, i.e., it is more likely that the true label is covered by the target labels. This is completely in line with the idea of data imprecisiation in the context of so-called superset learning (Lienen and Hüllermeier 2021b). Thus, ambiguating presumably corrupt training information appears to be a useful means to counter the influence of mislabeling.…”
Section: Credal Labelingsupporting
confidence: 82%
“…However, from a data modeling perspective, it is important to recognize that the imprecisiation of the target information pays off with higher validity, i.e., it is more likely that the true label is covered by the target labels. This is completely in line with the idea of data imprecisiation in the context of so-called superset learning (Lienen and Hüllermeier 2021b). Thus, ambiguating presumably corrupt training information appears to be a useful means to counter the influence of mislabeling.…”
Section: Credal Labelingsupporting
confidence: 82%
“…By contrast, data imprecisiation [24,34] refers to soft computing approaches by which data affected by some form of uncertainty are transformed into imprecise (soft) observations, that is distributions over possible instances, which are then used to train specialized ML algorithms. Formally speaking, an imprecisiation scheme is a function is : X × Y → [0, 1] X×Y , where X is the feature space.…”
Section: Data Augmentation and Imprecisiation Methods To Manage Insta...mentioning
confidence: 99%
“…More in particular, to provide a more self-contained and detailed discussion, we will focus our experimental analysis on a specific setting, the medical one, which is of particular relevance due to its critical characteristics as well as due to it being one of the fields of applications of ML in which the problem of IV has been more frequently acknowledged. Finally, the last part of the paper will aim to build on the rubble left by the first part, and it will focus on the hypothesis whether more advanced learning and regularization methods (grounding on, either, data augmentation [33] or data imprecisiation [34]) will achieve increased robustness in face of the same perturbations (H 2 ). To address these two research questions, and motivated by the lack of datasets that rep-resent and allow to investigate this complex form of uncertainty, we will rely on a large gold-standard medical dataset that had been proposed for the task of COVID-19 diagnosis, a major impactful concern, which was specifically constructed with the help of clinical laboratory medicine to study IV, grounding on previous knowledge in this domain [30,35,36].…”
Section: Introductionmentioning
confidence: 99%
“…Such methods can be found in a wide range of domains, including natural language processing [22] and computer vision [21,27,56,68]. As self-training can be considered as a general learning paradigm, where a model suggests itself labels to learn from, it has been wrapped around various model types, e.g., support vector machines [44], decision trees [62] and most prominently with neural networks [50]. Notably, it lays the foundation for so-called distillation models, e.g., in self-distillation [35] or studentteacher setups [53,68].…”
Section: Related Workmentioning
confidence: 99%