2019
DOI: 10.1109/access.2019.2901300
|View full text |Cite
|
Sign up to set email alerts
|

Multi-Agent Deep Reinforcement Learning for Multi-Object Tracker

Abstract: Multi-object tracking has been a key research subject in many computer vision applications. We propose a novel approach based on multi-agent deep reinforcement learning (MADRL) for multi-object tracking to solve the problems in the existing tracking methods, such as a varying number of targets, noncausal, and non-realtime. At first, we choose YOLO V3 to detect the objects included in each frame. Unsuitable candidates were screened out and the rest of detection results are regarded as multiple agents and formin… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
19
0

Year Published

2019
2019
2022
2022

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 42 publications
(19 citation statements)
references
References 37 publications
0
19
0
Order By: Relevance
“…To solve this problem, they used an auto-referee system and the occlusion problem by using CNN with the appearance feature, which also uses the spatial attention mechanism of insertion and location. Jiang et al proposed multi-agent deep reinforcement learning (MADRL) [138], which uses a learning method that uses Q-Learning (QL) to treat other agents as part of the current agent's environment. [56], which included a deformable attention to transformer encoder.…”
Section: Automatic Detection Learningmentioning
confidence: 99%
“…To solve this problem, they used an auto-referee system and the occlusion problem by using CNN with the appearance feature, which also uses the spatial attention mechanism of insertion and location. Jiang et al proposed multi-agent deep reinforcement learning (MADRL) [138], which uses a learning method that uses Q-Learning (QL) to treat other agents as part of the current agent's environment. [56], which included a deformable attention to transformer encoder.…”
Section: Automatic Detection Learningmentioning
confidence: 99%
“…This action, a t , produces an effect on the environment. This fact is observed by the interpreter who provides information to the agent about the new state, s t+1 , and the reward of the previous action, r t+1 , closing the loop [38,39]. Some authors consider that the interpreter is embedded in either the environment or the agent; in any case, the function of the interpreter is always present.…”
Section: Rl-inspired Controllermentioning
confidence: 99%
“…Our pre-trained human detector YOLOv3 [36,40,41] (Table 1) is configured to obtain for almost everyone (human) in the image (I i ) a bounding box that surrounds him. Considering a video V subdivided into n frames as…”
Section: Human Detector and Regions Of Interest (Rois)mentioning
confidence: 99%