1998
DOI: 10.1109/18.650998
|View full text |Cite
|
Sign up to set email alerts
|

Memory-universal prediction of stationary random processes

Abstract: We consider the problem of one-step-ahead prediction of a real-valued, stationary, strongly mixing random process fX i g 1 i=01. The best mean-square predictor of X 0 is its conditional mean given the entire infinite past fX i g 01 i=01. Given a sequence of observations X 1 X 2 1 1 1 X N , we propose estimators for the conditional mean based on sequences of parametric models of increasing memory and of increasing dimension, for example, neural networks and Legendre polynomials. The proposed estimators select b… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

5
53
1

Year Published

1998
1998
2016
2016

Publication Types

Select...
4
3

Relationship

0
7

Authors

Journals

citations
Cited by 57 publications
(59 citation statements)
references
References 55 publications
5
53
1
Order By: Relevance
“…Finally, the optimal balance between the complexity of the model and the memory size used for prediction must be determined. We observe that our results bear strong affinities to the approach taken by Modha and Masry (1998), while deviating from them in scope and methodology; see Remark 7 in Section 6 for a detailed comparison. An additional related work is the one by Campi and Kumar (1998), which deals with the problem of learning dynamical systems in a stationary environment.…”
Section: Introductionmentioning
confidence: 83%
See 1 more Smart Citation
“…Finally, the optimal balance between the complexity of the model and the memory size used for prediction must be determined. We observe that our results bear strong affinities to the approach taken by Modha and Masry (1998), while deviating from them in scope and methodology; see Remark 7 in Section 6 for a detailed comparison. An additional related work is the one by Campi and Kumar (1998), which deals with the problem of learning dynamical systems in a stationary environment.…”
Section: Introductionmentioning
confidence: 83%
“…One approach to incorporating temporal structure in order to form better predictors, by more appropriate complexity regularization, is described in this work. In particular, the optimal memory size that should be used in order to form a predictor is in principle derivable from the procedure (see also (Modha & Masry, 1998)), given information about the mixing nature of the time series (see Section 4 for a definition of mixing). It is thus hoped that many of the successful Machine Learning approaches to modeling static data will be extended to time series, with the benefit of a solid mathematical framework.…”
Section: Introductionmentioning
confidence: 99%
“…Implicitly, while establishing the rates of convergence results in Theorem 2.1 and Corollary 2.1, we assumed that these least-squares estimators can indeed be computed. Such an assumption is the very basis for applying Vapnik's empirical risk minimization theory to neural networks, and has been widely used in the literature dealing with rates of convergence results for neural networks and other models, see, for example, Barron (1994), Barron, Birgé, & Massart (1996), Breiman (1993), Haussler (1992), Kearns (1997), Lugosi & Nobel (1995), Lugosi & Zeger (1996), McCaffrey & Gallant (1994), Modha & Masry (1996, 1998, Vapnik (1982Vapnik ( , 1995, and White (1989).…”
Section: Remark 23 (Computational Complexity Of Nonlinear Least-squamentioning
confidence: 99%
“…In this case, it is known that the complexityregularized regression estimator is order-universal, that is, the complexity-regularized regression estimator, which does not know the true dimension m , delivers the same rate of integrated mean-squared error as that delivered by an estimator that knows the true dimension. See Barron (1994), Barron, Birgé, & Massart (1996), and Modha & Masry (1998). Establishing order-universality results for prequential and cross-validated regression estimators currently remains an open problem.…”
Section: Remark 24 (Comparison With Complexity-regularized Least Squmentioning
confidence: 99%
See 1 more Smart Citation