Smoothing, Filtering and Prediction - Estimating the Past, Present and Future 2012
DOI: 10.5772/39257
|View full text |Cite
|
Sign up to set email alerts
|

Robust Prediction, Filtering and Smoothing

Abstract: Unit 405, Office Block, Hotel Equatorial Shanghai This book describes the classical smoothing, filtering and prediction techniques together with some more recently developed embellishments for improving performance within applications. It aims to present the subject in an accessible way, so that it can serve as a practical guide for undergraduates and newcomers to the field. The material is organised as a ten-lecture course. The foundations are laid in Chapters 1 and 2, which explain minimum-mean-square-error … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
15
0

Year Published

2014
2014
2022
2022

Publication Types

Select...
5
1

Relationship

0
6

Authors

Journals

citations
Cited by 9 publications
(15 citation statements)
references
References 22 publications
0
15
0
Order By: Relevance
“…This bears a striking similarity to the optimal combination rule for two inputs (Ernst and Bülthoff 2004), which is a Kalman filter (Kalman and Bucy 1961). The static filter gain ( L ) is given by L=true⌊σA2PAσA2σB2+σA2PA+σB2PBσB2PBσB2σA2+σB2PB+σA2PAtrue⌋, where P A is the response of the channel (or sensor) tuned to input A , and σ A 2 is the variance, with terms bearing the subscript B corresponding to a second channel (Einicke 2012). Hence, the filter weights each input by the inverse of its contribution to the total variance.…”
Section: Discussionmentioning
confidence: 99%
“…This bears a striking similarity to the optimal combination rule for two inputs (Ernst and Bülthoff 2004), which is a Kalman filter (Kalman and Bucy 1961). The static filter gain ( L ) is given by L=true⌊σA2PAσA2σB2+σA2PA+σB2PBσB2PBσB2σA2+σB2PB+σA2PAtrue⌋, where P A is the response of the channel (or sensor) tuned to input A , and σ A 2 is the variance, with terms bearing the subscript B corresponding to a second channel (Einicke 2012). Hence, the filter weights each input by the inverse of its contribution to the total variance.…”
Section: Discussionmentioning
confidence: 99%
“…This so‐called “state” (or “transition,” or “system”) equation describes the continuous evolution of some “hidden” (or “latent”) states S which, being now stochastic, directly account for modeling errors. This vector of usually unmeasurable variables can be estimated from the measured outputs [ Einicke , ]. bold-italicσ is called “diffusion term,” “state noise,” or “level disturbance” and accounts for modeling errors by making the states uncertain or random.…”
Section: Methodsmentioning
confidence: 97%
“…Forecasting denotes the generation of model outputs (and states) starting from the last observation up to an arbitrary number of time steps in the future. This process is also loosely described as making predictions (in the validation period) [ Dietzel and Reichert , ; Renard et al ., ; Law and Stuart , ; Einicke , ], simulating [ Platen and Bruti‐Liberati , ] or, more precisely, ex‐post hindcasting (when the input is assumed to be known) [ Beven and Young , ].…”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…Fig. 1 shows the estimate of the correlation, i.e., z k , using the proposed filter, as well as the results of naive application of the Extended Kalman Filter (EKF) based on linearization of (2) [10], [11]. Clearly, state estimation using proposed optimal filter is closest to the true state of the system.…”
Section: Model Formulationmentioning
confidence: 99%