Proceedings of the 1st International Workshop on Reinforcement Learning for Energy Management in Buildings &Amp; Cities 2020
DOI: 10.1145/3427773.3427870
|View full text |Cite
|
Sign up to set email alerts
|

Benchmarking Multi-Agent Deep Reinforcement Learning Algorithms on a Building Energy Demand Coordination Task

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

1
12
1

Year Published

2021
2021
2023
2023

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 10 publications
(16 citation statements)
references
References 3 publications
1
12
1
Order By: Relevance
“…To compare our results with those of previous publications, we made sure that the scenarios and scenario parameters matched those of Papoudakis et al [5] and Atrazhev et al [19], and the results were compared to the results of those previous works.…”
Section: Methodsmentioning
confidence: 99%
See 3 more Smart Citations
“…To compare our results with those of previous publications, we made sure that the scenarios and scenario parameters matched those of Papoudakis et al [5] and Atrazhev et al [19], and the results were compared to the results of those previous works.…”
Section: Methodsmentioning
confidence: 99%
“…To remain consistent with previous publications, the LBF scenarios selected for this study are 8x8-2p-2f-coop, 2s-8x8-2p-2f-coop, 10x10-3p-3f, and 2s-10x10-3p-3f. Algorithms are also selected based on these criteria: IQL [16], IA2C [6], IPPO [17], MAA2C [5], MAPPO [7], VDN [10] and QMIX [9] were selected as they are studied in both Papoudakis et al [5] and in Atrazhev et al [19] and represent an acceptable assortment of independent algorithms, centralized critic CLDE algorithms, and value factorization CLDE algorithms.…”
Section: Methodsmentioning
confidence: 99%
See 2 more Smart Citations
“…The CityLearn environment is an OpenAI environment which allows the control of domestic hot water and chilled water storage in a district environment. This environment was also used by Dhamankar et al [31] who compared three classes of multi-agent RL algorithms for demand response and coordination of a district. Note that for this environ-ment, the conditioning loads were precomputed for a given set of climates.…”
Section: Action a T Reward R T+1mentioning
confidence: 99%