2020
DOI: 10.1007/s00138-020-01090-5
|View full text |Cite
|
Sign up to set email alerts
|

Boosting binary masks for multi-domain learning through affine transformations

Abstract: In this work, we present a new, algorithm for multi-domain learning. Given a pretrained architecture and a set of visual domains received sequentially, the goal of multi-domain learning is to produce a single model performing a task in all the domains together. Recent works showed how we can address this problem by masking the internal weights of a given original conv-net through learned binary variables. In this work, we provide a general formulation of binary mask based models for multi-domain learning by af… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
3
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
3
1

Relationship

1
3

Authors

Journals

citations
Cited by 4 publications
(3 citation statements)
references
References 60 publications
(156 reference statements)
0
3
0
Order By: Relevance
“…In test time, the learned binary mask is multiplied by the weights of the convolutional layer. Expanding on this idea, Mancini et al [7,8] also makes use of masks, however, they learn an affine transformation of the weights through the use of the mask and some extra parameters. Focusing on increasing the accuracy with masks, Chattopadhyay et al [2] proposes a soft-overlap loss to encourage the masks to be domain-specific by minimizing the overlap between them.…”
Section: Intersection Between Masksmentioning
confidence: 99%
“…In test time, the learned binary mask is multiplied by the weights of the convolutional layer. Expanding on this idea, Mancini et al [7,8] also makes use of masks, however, they learn an affine transformation of the weights through the use of the mask and some extra parameters. Focusing on increasing the accuracy with masks, Chattopadhyay et al [2] proposes a soft-overlap loss to encourage the masks to be domain-specific by minimizing the overlap between them.…”
Section: Intersection Between Masksmentioning
confidence: 99%
“…Since our model can identify data clusters through the previously described procedure, we can design a way to specialize the function f θ to each domain. Inspired by multi-domain learning [36,39,37,27,29], we can achieve this with domain-specific components. For simplicity, let us consider the parameters θ to be split into two sets, i.e.…”
Section: Cluster-specific Modelsmentioning
confidence: 99%
“…Note that θ s is actually a set θ s = {θ d s } D d=1 where θ d s are the parameters specific to the d-th domain. To tailor the model to a specific domain, we can consider multiple ways to include θ s , such as direct influence on the agnostic parameters θ a [39,27,29] or residual activations [36,37]. Here we follow the latter strategy, since the former relies on the robustness of θ a , which is harder to guarantee in FL.…”
Section: Cluster-specific Modelsmentioning
confidence: 99%