k Baseline right-parenthesis comma EndLayout"/>
where , , and denote the state, the input, and the output vectors, respectively. Compared with the deterministic nonlinear discrete‐time model in (2.78) and (2.79), two additional variables are included in the above stochastic model, which are the process noise, , and the measurement noise, . These two random variables take account of model inaccuracies and other sources of uncertainty. The more accurate the model is, the smaller the contribution of noise terms will be. These two noise sequences are assumed to be white, independent of each other, and independent from the initial state. The probabilistic model of the state evolution in (4.3) is assumed to be a first‐order Markov process, and therefore, can be rewritten as the following state‐transition probability density function (PDF) [46]:
(4.5)
Similarly, the measurement model in (4.4) can be represented by the following PDF:
(4.6)
The input sequence and the available measurement sequence at time instant are denoted by and , respectively. These two sequences form the available information at time , hence the union of these two sets is called the information set, [47].
A filter uses the inputs and available observations up to time instant , to estimate the state at , . In other words, a filter tries to solve an inverse problem to infer the states (cause) from the observations (effect). Due to uncertainties, different values of the state could have led to the obtained measurement sequence, . The Bayesian framework allows us to associate a degree of belief to these possibly valid values of state. The main idea here is to start from an initial density for the state vector, , and recursively calculate the posterior PDF, based on the measurements. This can be done by a filtering algorithm that includes two‐stages of prediction and update [46].
In the prediction stage, the Chapman–Kolmogorov equation can be used to calculate the prediction density, , from , which is provided by the update stage of the previous iteration. Using (4.5), the prediction density can be computed as [46, 47]:
(4.7)
When a new measurement is obtained, the prediction stage is followed by the update stage, where the above prediction density will play the role of the prior. Bayes' rule is used to compute the posterior density of the state as [46, 47]:
(4.8)