Proceedings of the 25th International Conference on Machine Learning - ICML '08 2008
DOI: 10.1145/1390156.1390186
|View full text |Cite
|
Sign up to set email alerts
|

Learning from incomplete data with infinite imputations

Abstract: We address the problem of learning decision functions from training data in which some attribute values are unobserved. This problem can arise, for instance, when training data is aggregated from multiple sources, and some sources record only a subset of attributes. We derive a generic joint optimization problem in which the distribution governing the missing values is a free parameter. We show that the optimal solution concentrates the density mass on finitely many imputations, and provide a corresponding alg… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

1
13
0

Year Published

2009
2009
2021
2021

Publication Types

Select...
5
3
2

Relationship

0
10

Authors

Journals

citations
Cited by 31 publications
(14 citation statements)
references
References 6 publications
1
13
0
Order By: Relevance
“…Building upon this generative model strategy, some approaches have considered integrating out the missing values based on a simple logistic regression function [6], [7]. Other versions of this approach proposed an explicit simultaneous learning of the model and the decision function [8], [9]. While probabilistic generative models provided a nice and elegant approach to the incomplete-features problem showing good results on small datasets, they are not suitable for many modern machine learning applications because:…”
Section: A Related Workmentioning
confidence: 99%
“…Building upon this generative model strategy, some approaches have considered integrating out the missing values based on a simple logistic regression function [6], [7]. Other versions of this approach proposed an explicit simultaneous learning of the model and the decision function [8], [9]. While probabilistic generative models provided a nice and elegant approach to the incomplete-features problem showing good results on small datasets, they are not suitable for many modern machine learning applications because:…”
Section: A Related Workmentioning
confidence: 99%
“…However, this method may miss some important information, and several works have demonstrated the dangers of simply removing cases using the case deletion [6,7]. The next one is to learn without handing of missing values such as learning Bayesian Networks [8], Artificial Neural Networks [9,10] and missing values fish-net learning algorithms [11].…”
Section: A Related Workmentioning
confidence: 99%
“…In the first approach, case deletion [8] is a method which is used basically to ignore those cases that come along with the missing values and to complete the learning progression only to utilize the residual instances. For the second technique of the learning without handling of missing data schemes, such as Bayesian Networks method, Artificial Neural Networks method, and some more approaches are demonstrated in [9], [10]. The third approach is entirely dissimilar to the above two techniques.…”
Section: II Related Workmentioning
confidence: 99%