2019
DOI: 10.48550/arxiv.1909.11944
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Multiple Object Forecasting: Predicting Future Object Locations in Diverse Environments

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
5
0

Year Published

2020
2020
2020
2020

Publication Types

Select...
2

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(5 citation statements)
references
References 0 publications
0
5
0
Order By: Relevance
“…Notably, the Stanford Drone Dataset (SDD) [50] is used in many works [53,13,30] for trajectory prediction with drone videos. Other works have also looked into pedestrian prediction in dashcam videos [42,57,25,28] and first-person videos [70,58]. Many vehicle trajectory datasets [6,8,71] have been proposed as a result of self-driving's surging popularity.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Notably, the Stanford Drone Dataset (SDD) [50] is used in many works [53,13,30] for trajectory prediction with drone videos. Other works have also looked into pedestrian prediction in dashcam videos [42,57,25,28] and first-person videos [70,58]. Many vehicle trajectory datasets [6,8,71] have been proposed as a result of self-driving's surging popularity.…”
Section: Related Workmentioning
confidence: 99%
“…A notable bottleneck for existing works is that the current model is closely coupled with the video cameras on which it is trained, and generalizes poorly on new cameras with novel views or scenes. For example, prior works have proposed various models to forecast a pedestrian's trajectories in video cameras of different types such as stationary outdoor cameras [44,32,1,19,29,37], drone cameras [53,13,30], ground-level egocentric cameras [70,47,58], or dash cameras [42,57,8]. Models that rely on RGB pixel inputs are especially vulnerable to view changes [34,33,53].…”
Section: Introductionmentioning
confidence: 99%
“…TraPHic [10] exploits the interaction between nearby heterogeneous objects. DTP [49] and STED [48] use encoder-decoder schemes using optical flow and past locations and scales of the objects. Yao et al [63] added the planned egomotion to further improve the prediction.…”
Section: Related Workmentioning
confidence: 99%
“…STED [48]. STED is a spatial-temporal encoderdecoder that models visual features by optical flow and temporal features by the past bounding boxes through GRU encoders.…”
Section: Baselinesmentioning
confidence: 99%
See 1 more Smart Citation