2020
DOI: 10.1007/978-3-030-58555-6_46
|View full text |Cite
|
Sign up to set email alerts
|

WeightNet: Revisiting the Design Space of Weight Networks

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
44
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
3
3
3

Relationship

0
9

Authors

Journals

citations
Cited by 83 publications
(53 citation statements)
references
References 24 publications
0
44
0
Order By: Relevance
“…The Meta networks (Munkhdalai and Yu, 2017) combine the parameters predicted by meta learner and the weights trained cross tasks making rapid generalization. The WeightNet (Ma et al, 2020) unifies the SENet (Hu et al, 2018) and CondConv (Yang et al, 2019), trains the network in the kernel space and takes the network to generate weights for the classification task.…”
Section: Meta Learning For Parameters Generalizationmentioning
confidence: 99%
“…The Meta networks (Munkhdalai and Yu, 2017) combine the parameters predicted by meta learner and the weights trained cross tasks making rapid generalization. The WeightNet (Ma et al, 2020) unifies the SENet (Hu et al, 2018) and CondConv (Yang et al, 2019), trains the network in the kernel space and takes the network to generate weights for the classification task.…”
Section: Meta Learning For Parameters Generalizationmentioning
confidence: 99%
“…To solve this problem, dynamic filters are proposed one after another. One kind of dynamic filter [10,51,78,88] predicts coefficients to combine several expert filters which are then shared across all spatial pixels. Another kind of dynamic filter [7,39,42,58,66,67,83,92] generates spatial-specific filters.…”
Section: Related Workmentioning
confidence: 99%
“…Several recent works [40], [37], [41], [42] focus on "input dependent" dynamic inference, i.e., the path to inference is determined by the individual input, where the "easy" inputs go through simple paths and thus reduce computation on average. Our "dynamic rerouting" is fundamentally different in that ours is input independent, i.e., any input data can be switched from small to tiny sub-models from any designated switching point (e.g., 4 points for ResNet50).…”
Section: Comparison To Existing Dynamic Dnnsmentioning
confidence: 99%
“…Slimmable neural network [37] has one result that achieves 1.4% less accuracy than our tiny ResNet50 with 5× parameters (27.06% vs 5.43%). CondConv [41] or WeightNet [42] are even larger than the dense ResNet50.…”
Section: Comparison To Existing Dynamic Dnnsmentioning
confidence: 99%