2019
DOI: 10.1109/lcomm.2019.2922658
|View full text |Cite
|
Sign up to set email alerts
|

Optimizing Pipelined Computation and Communication for Latency-Constrained Edge Learning

Abstract: Consider a device that is connected to an edge processor via a communication channel. The device holds local data that is to be offloaded to the edge processor so as to train a machine learning model, e.g., for regression or classification. Transmission of the data to the learning processor, as well as training based on Stochastic Gradient Descent (SGD), must be both completed within a time limit. Assuming that communication and computation can be pipelined, this letter investigates the optimal choice for the … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
13
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
4
2

Relationship

0
6

Authors

Journals

citations
Cited by 16 publications
(13 citation statements)
references
References 9 publications
0
13
0
Order By: Relevance
“…Popular traditional methods for distributed inference in the online learning context use incremental updates among nodes [1]- [3], [36]- [38]. In recent years, other stochastic gradient-descent (SGD) based methods were developed for federated learning [36], [37]. While these methods do not require prior knowledge of the observation distributions, they use orthogonal channels among node transmissions, which results in high bandwidth and energy consumption.…”
Section: Related Workmentioning
confidence: 99%
“…Popular traditional methods for distributed inference in the online learning context use incremental updates among nodes [1]- [3], [36]- [38]. In recent years, other stochastic gradient-descent (SGD) based methods were developed for federated learning [36], [37]. While these methods do not require prior knowledge of the observation distributions, they use orthogonal channels among node transmissions, which results in high bandwidth and energy consumption.…”
Section: Related Workmentioning
confidence: 99%
“…Other recent general surveys can be found in [21]- [23]. Going into more specific contributions, the authors of [24] consider an edge machine learning system, where an edge processor runs an algorithm based on Stochastic Gradient Descent (SGD). In particular, they investigate the trade-off between latency and accuracy, by optimizing the packet payload size, given the overhead of each data packet transmission and the ratio between the computation and communication rates.…”
Section: A Related Workmentioning
confidence: 99%
“…Using (24), the evolution of the virtual queue Y k (t) can be written as (14), replacing the closed form expression G k (t) with G k (t). Using G k (t) is useful for the virtual queue's update, but it is not directly related to the number of quantization bits, which affect the learning accuracy.…”
Section: Data-driven Control Of Learning Accuracymentioning
confidence: 99%
See 1 more Smart Citation
“…In the edge learning framework, having in mind the goal of learning and the resources dedicated to that goal, several trade-offs are possible, like the trade-off between power consumption and delay, between accuracy and delay, etc. The authors of [61] consider an edge machine learning system, where an edge processor runs an algorithm based on stochastic gradient descent (SGD), to reach a trade-off between latency and accuracy, by optimizing the packet payload size, given the overhead of each data packet transmission and the ratio between the computation and the communication rates. In [62], the authors proposed an algorithm to maximize the learning accuracy under latency constraints.…”
Section: Goal-oriented Communicationmentioning
confidence: 99%