2019 IEEE 13th International Conference on Self-Adaptive and Self-Organizing Systems (SASO) 2019
DOI: 10.1109/saso.2019.00015
|View full text |Cite
|
Sign up to set email alerts
|

Autonomous Management of Energy-Harvesting IoT Nodes Using Deep Reinforcement Learning

Abstract: Reinforcement learning (RL) is capable of managing wireless, energy-harvesting IoT nodes by solving the problem of autonomous management in non-stationary, resource-constrained settings. We show that the state-of-the-art policy-gradient approaches to RL are appropriate for the IoT domain and that they outperform previous approaches. Due to the ability to model continuous observation and action spaces, as well as improved function approximation capability, the new approaches are able to solve harder problems, p… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
18
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
7
2

Relationship

2
7

Authors

Journals

citations
Cited by 21 publications
(18 citation statements)
references
References 17 publications
0
18
0
Order By: Relevance
“…It employs a reward function and learns through interaction of an agent with its environment, with no need for a complete control model or explicit supervision [22]. An RL agent is trained to improve a task by learning from experience, that is, interacting with that particular task in context [62]. The algorithm is trained with the goal to maximize the cumulative reward.…”
Section: ) Machine Learning (Ml)mentioning
confidence: 99%
See 1 more Smart Citation
“…It employs a reward function and learns through interaction of an agent with its environment, with no need for a complete control model or explicit supervision [22]. An RL agent is trained to improve a task by learning from experience, that is, interacting with that particular task in context [62]. The algorithm is trained with the goal to maximize the cumulative reward.…”
Section: ) Machine Learning (Ml)mentioning
confidence: 99%
“…An example can be seen in [37], where Edalat et al use reinforcement learning for network lifetime optimization. Challenges with RL are the design of the reward function, as this requires an in-depth knowledge of the domain and the system goals, as well as a potentially high training effort [62].…”
Section: ) Machine Learning (Ml)mentioning
confidence: 99%
“…For example, in an energy-harvesting management system, PPO algorithm [292] is used to control IoT nodes for power allocation. The action space, as stated in [245], is sampled from a Gaussian distribution to denote the load of each node ranging from 0% to 100%. Similarly, in another work [9] that studied energy harvesting WSNs, the Actor-Critic [179] algorithm is implemented to control the packet rate during transmission.…”
Section: Rl Categorizationmentioning
confidence: 99%
“…It can successfully control more than 20 kinds of physics tasks such as cart-pole swing-up, legged locomotion, car driving, and Reacher domain with multiple continuous action spaces (Lillicrap et al, 2016). In engineering, DRL has been widely used in optimization and control problems in practical applications such as robotics (Gu et al, 2017), HAVC control (Chen et al, 2018), and energy harvesting (Long & Büyüköztürk, 2020;Murad et al, 2019). These successes demonstrate the ability of DRL to learn complex tasks that require expert-level knowledge and experience.…”
Section: Introductionmentioning
confidence: 99%