2022
DOI: 10.1609/aaai.v36i7.20778
|View full text |Cite
|
Sign up to set email alerts
|

DISTREAL: Distributed Resource-Aware Learning in Heterogeneous Systems

Abstract: We study the problem of distributed training of neural networks (NNs) on devices with heterogeneous, limited, and time-varying availability of computational resources. We present an adaptive, resource-aware, on-device learning mechanism, DISTREAL, which is able to fully and efficiently utilize the available resources on devices in a distributed manner, increasing the convergence speed. This is achieved with a dropout mechanism that dynamically adjusts the computational complexity of training an NN by randomly … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
5
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
3
2

Relationship

0
5

Authors

Journals

citations
Cited by 6 publications
(5 citation statements)
references
References 33 publications
0
5
0
Order By: Relevance
“…Masks are updated every round. In DISTREAL [80], the authors explore how subsets can be trained in environments with time-varying computational resources that change faster than FL rounds and are not known in advance. A mini-batch level granularity for training subsets by randomly switching filters of the CNN during training is achieved.…”
Section: Nn Architecture Heterogeneity Based On Fedavgmentioning
confidence: 99%
See 3 more Smart Citations
“…Masks are updated every round. In DISTREAL [80], the authors explore how subsets can be trained in environments with time-varying computational resources that change faster than FL rounds and are not known in advance. A mini-batch level granularity for training subsets by randomly switching filters of the CNN during training is achieved.…”
Section: Nn Architecture Heterogeneity Based On Fedavgmentioning
confidence: 99%
“…The attributes scale and granularity are often neglected, are hidden behind the technique, and lack discussion in the papers. The reported scale in the resources supported by the techniques ranges from 4× − 25× [12,41,52,61,71,77,79,80,85,101] up to 100× − 250× [25,87], yet it remains unclear whether training at such high scales is still effective. Hence, while all approaches show the effectiveness of their solution in certain scenarios, it often remains unclear whether devices with low resources or stale devices can make a meaningful contribution that advances the global model.…”
Section: Open Problems and Future Directionsmentioning
confidence: 99%
See 2 more Smart Citations
“…Other solutions entail work from the clients' side. Under a known time limitations, workers either speed up local training by allocating more computing resources or training models partially [27] with gradient sparsification [28], or reduce the communication time by compressing the model [29]. Such methods do not apply in those conditions in which vehicles completely lose the connection with the server.…”
Section: Challenges Of Federated Learning In V2xmentioning
confidence: 99%