2023
DOI: 10.1109/tnnls.2021.3089493
|View full text |Cite
|
Sign up to set email alerts
|

SMIX(λ): Enhancing Centralized Value Functions for Cooperative Multiagent Reinforcement Learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2023
2023
2025
2025

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 13 publications
(2 citation statements)
references
References 16 publications
0
2
0
Order By: Relevance
“…1) Selection of Value Decomposition Network: After the advent of QMIX, many derivative algorithms [32], [34], [35] are based on the hyper-network to establish the connection between joint action-values and individual action values. However, in MCVD, the joint action-values is obtained by summing all individual action values, just like VDN does.…”
Section: Details Of Mcvdmentioning
confidence: 99%
“…1) Selection of Value Decomposition Network: After the advent of QMIX, many derivative algorithms [32], [34], [35] are based on the hyper-network to establish the connection between joint action-values and individual action values. However, in MCVD, the joint action-values is obtained by summing all individual action values, just like VDN does.…”
Section: Details Of Mcvdmentioning
confidence: 99%
“…Hausknecht et al [17] proposed the DRQN algorithm, which replaces the fully connected layer of DQN with recurrent neural network and can better handle the missing information.In 2016, Wang et al [18] proposed the Dueling-DQN algorithm to optimize the internal of DQN framework and better results in performance evaluation. In the same year, DeepMind proposed the AlphaGo algorithm [19], which mainly utilizes Monte Carlo tree search and deep learning techniques, and then learns Go games through human prior knowledge, and soon AlphaGo defeated the Go master Lee Sedol, attracting widespread attention from the world.…”
Section: Introductionmentioning
confidence: 99%