2017
DOI: 10.1007/978-3-319-71246-8_13
|View full text |Cite
|
Sign up to set email alerts
|

PAC-Bayesian Analysis for a Two-Step Hierarchical Multiview Learning Approach

Abstract: We study a two-level multiview learning with more than two views under the PAC-Bayesian framework. This approach, sometimes referred as late fusion, consists in learning sequentially multiple view-specific classifiers at the first level, and then combining these view-specific classifiers at the second level. Our main theoretical result is a generalization bound on the risk of the majority vote which exhibits a term of diversity in the predictions of the view-specific classifiers. From this result it comes out … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
16
0

Year Published

2017
2017
2022
2022

Publication Types

Select...
3
3

Relationship

4
2

Authors

Journals

citations
Cited by 9 publications
(18 citation statements)
references
References 20 publications
0
16
0
Order By: Relevance
“…The question of finding a relevant prior in a domain adaptation situation is an exciting direction which could also be exploited when some few target labels are available. Moreover, this notion of prior distribution could model information learned from previous tasks as pointed out by Pentina and Lampert [61], or from other views/representations of the data [62]. This suggests that we can extend our analyses to multisource domain adaptation [63,64,27,65] and lifelong learning where the objective is to perform well on future tasks, for which no data has been observed so far [66].…”
Section: Discussionmentioning
confidence: 87%
“…The question of finding a relevant prior in a domain adaptation situation is an exciting direction which could also be exploited when some few target labels are available. Moreover, this notion of prior distribution could model information learned from previous tasks as pointed out by Pentina and Lampert [61], or from other views/representations of the data [62]. This suggests that we can extend our analyses to multisource domain adaptation [63,64,27,65] and lifelong learning where the objective is to perform well on future tasks, for which no data has been observed so far [66].…”
Section: Discussionmentioning
confidence: 87%
“…In this section, we give a general multiview PAC-Bayesian theorem Goyal et al [2017] that takes the form of a generalization bound for the Gibbs risk in the context of a two-level hierarchy of distributions. A key step in PAC-Bayesian proofs is the use of a change of measure inequality [McAllester, 2003], based on the Donsker-Varadhan inequality [Donsker and Varadhan, 1975].…”
Section: The General Multiview Pac-bayesian Theoremmentioning
confidence: 99%
“…Since the classes are unbalanced, we report the accuracy along with F1-measure for the methods and all the scores are averaged over all the one-vs-all classification problems. We consider two multiview learning algorithms based on our two-step hierarchical strategy, and compare the PB-MVBoost algorithm described in Section 4, with a previously developed multiview learning algorithm Goyal et al [2017], based on classifier late fusion approach Snoek et al [2005], and referred to as Fusion all Cq . Concretely, at the first level, this algorithm trains different view-specific linear SVM models with different hyperparameter C values (12 values between 10 −8 and 10 3 ).…”
Section: Experimental Protocolmentioning
confidence: 99%
See 1 more Smart Citation
“…Basically, they train multiple view-specific models either by alternatively optimizing them, "teaching" one another, or by fostering their smoothness in predictions. The final step of such techniques consists in aggregating the predictions of the view-specific classifiers, for instance by majority vote [21,11] or by weighted majority vote [19,13]. Note that these methods usually face the following issues: their performances are degraded by the computational overload of training and testing multiple learners; also, by usually making the assumption that the views' common information is the only worth keeping, they boil down to denoising the single views from their uncorrelated information.…”
Section: Related Workmentioning
confidence: 99%