2017 IEEE International Conference on Computer Vision (ICCV) 2017
DOI: 10.1109/iccv.2017.43
|View full text |Cite
|
Sign up to set email alerts
|

Tracking as Online Decision-Making: Learning a Policy from Streaming Videos with Reinforcement Learning

Abstract: We formulate tracking as an online decision-making process, where a tracking agent must follow an object despite ambiguous image frames and a limited computational budget. Crucially, the agent must decide where to look in the upcoming frames, when to reinitialize because it believes the target has been lost, and when to update its appearance model for the tracked object. Such decisions are typically made heuristically. Instead, we propose to learn an optimal decision-making policy by formulating tracking as a … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
48
0

Year Published

2018
2018
2020
2020

Publication Types

Select...
5
4

Relationship

0
9

Authors

Journals

citations
Cited by 103 publications
(48 citation statements)
references
References 52 publications
(124 reference statements)
0
48
0
Order By: Relevance
“…The training of the proposed method is stable, and the agents obtain a good strategy immediately because all the agents share the parameters and the gradients are averaged as shown in Eqs. (14) and (16). In other words, the agents can obtain N (> 10 5 ) experience at one backward pass in contrast to usual reinforcement learning where only one experience is obtained at one backward pass.…”
Section: Applications and Resultsmentioning
confidence: 98%
See 1 more Smart Citation
“…The training of the proposed method is stable, and the agents obtain a good strategy immediately because all the agents share the parameters and the gradients are averaged as shown in Eqs. (14) and (16). In other words, the agents can obtain N (> 10 5 ) experience at one backward pass in contrast to usual reinforcement learning where only one experience is obtained at one backward pass.…”
Section: Applications and Resultsmentioning
confidence: 98%
“…The fundamental algorithm of the proposed method and the experimental results of the image denoising, image restoration, and local color enhancement have already been presented in our preliminary study [6]. This paper provides a more detailed explanation and a new application of the pixelRL to saliency-arXiv:1912.07190v1 [cs.CV] 16 Dec 2019 driven image editing.…”
Section: Introductionmentioning
confidence: 99%
“…Recently, in the computer vision community there are also preliminary attempts of applying deep RL to traditional tasks, e.g., object localization [37] and region proposal [38]. There are also methods of visual tracking relying on RL [39], [40], [41], [42]. However, they are distinct from our work, as they formulate passive tracking with RL and do not consider camera controls.…”
Section: Reinforcement Learningmentioning
confidence: 99%
“…Employing the deep RL algorithms into computer vision problems could benefit from the experience. In fact, RL has been studied for visual tracking in several recent works [27], [28], [43], [44]. Huang et al succeed in utilizing Q-learning [29] for shallow-level or high-level feature selection [27].…”
Section: Deep Reinforcement Learningmentioning
confidence: 99%