In contrast to the widely studied problem of recognizing an action given a complete sequence, action anticipation aims to identify the action from only partially available videos. As such, it is therefore key to the success of computer vision applications requiring to react as early as possible, such as autonomous navigation. In this paper, we propose a new action anticipation method that achieves high prediction accuracy even in the presence of a very small percentage of a video sequence. To this end, we develop a multi-stage LSTM architecture that leverages context-aware and action-aware features, and introduce a novel loss function that encourages the model to predict the correct class as early as possible. Our experiments on standard benchmark datasets evidence the benefits of our approach; We outperform the state-of-the-art action anticipation methods for early prediction by a relative increase in accuracy of 22.0% on JHMDB-21, 14.0% on UT-Interaction and 49.9% on UCF-101.
This paper presents a comparative analysis of different pedestrian dataset characteristics. The main goal of the research is to determine what characteristics are desirable for improved training and validation of pedestrian detectors and classifiers. The work focuses on those aspects of the dataset which affect classification success using the most common boosting methods.Dataset characteristics such as image size, aspect ratio, geometric variance and the relative scale of positive class instances (pedestrians) within the training window form an integral part of classification success. This paper will examine the effects of varying these dataset characteristics with a view to determining the recommended attributes of a high quality and challenging dataset. While the primary focus is on characteristics of the positive training dataset, some discussion of desirable attributes for the negative dataset is important and is therefore included. This paper also serves to publish our current pedestrian dataset in various forms for non-commercial use by the scientific community. We believe the published dataset to be one of the largest, most flexible, and representative datasets available for pedestrian/person detection tasks.
Abstract-This paper presents a Weak Classifier that is extremely fast to compute, yet highly discriminant. This Weak Classifier may be used in, for example, a boosting framework and is the result of a novel way of organizing and evaluating Histograms of Oriented Gradients. The method requires only one access to main memory to evaluate each feature, in comparison with the more well-known Haar features which require somewhere between six and nine memory accesses to evaluate each feature. This low memory bandwidth makes the Weak Classifier especially ideal for use in small systems with little or no memory cache available.The presented Weak Classifier has been extensively tested in a boosted framework on data sets consisting of pedestrians and various road signs. The classifier yields detection results that are far superior than the results obtained from Haar features when tested on road signs and similar structures, whereas the detection results are comparable to those of Haar features when tested on pedestrians. In addition, the computational resources necessary for these results have been shown to be considerably smaller for the new weak classifier.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
customersupport@researchsolutions.com
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.