2014
DOI: 10.1016/j.imavis.2014.08.009
|View full text |Cite
|
Sign up to set email alerts
|

Adaptive visual tracking using the prioritized Q-learning algorithm: MDP-based parameter learning approach

Abstract: a b s t r a c tThis paper introduces an adaptive visual tracking method that combines the adaptive appearance model and the optimization capability of the Markov decision process. Most tracking algorithms are limited due to variations in object appearance from changes in illumination, viewing angle, object scale, and object shape. This paper is motivated by the fact that tracking performance degradation is caused not only by changes in object appearance but also by the inflexible controls of tracker parameters… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2017
2017
2021
2021

Publication Types

Select...
3
1

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(2 citation statements)
references
References 47 publications
(76 reference statements)
0
2
0
Order By: Relevance
“…The resulting partially-observable MDP (POMDP) violates Markov independence assumptions: actions depend on the entire history of observations rather than just the current one [20,47]. As in [16,23], we account for this partial observability by maintaining a memory that captures beliefs about the world, which we update over time (Sec. 3).…”
Section: Contribution 1 (Interactive Video Processing)mentioning
confidence: 99%
See 1 more Smart Citation
“…The resulting partially-observable MDP (POMDP) violates Markov independence assumptions: actions depend on the entire history of observations rather than just the current one [20,47]. As in [16,23], we account for this partial observability by maintaining a memory that captures beliefs about the world, which we update over time (Sec. 3).…”
Section: Contribution 1 (Interactive Video Processing)mentioning
confidence: 99%
“…Interestingly, such an "end-to-end" learning philosophy is often embraced by the multi-object tracking community, where strategies for online reinitialization and data association are learned from data [26,62,31]. Most related to us are [62], who use an MDP for multi-object tracking, and [23], who use RL for single target tracking. Both works use heuristics to reduce policy learning to a supervised learning task, avoiding the need to reason about rewards in the far future.…”
Section: Related Workmentioning
confidence: 99%