Kushner equation

From Wikipedia, the free encyclopedia
Jump to: navigation, search

In filtering theory the Kushner equation[1] (after Harold Kushner) is an equation for the conditional probability density of the state of a stochastic non-linear dynamical system, given noisy measurements of the state. It therefore provides the solution of the nonlinear filtering problem in estimation theory. The equation is sometimes referred to as the Stratonovich–Kushner[2][3][4][5] (or Kushner–Stratonovich) equation. However, the correct equation in terms of Itō calculus was first derived by Kushner although a more heuristic Stratonovich version of it appeared already in Stratonovich's works in late fifties. However, the derivation in terms of Itō calculus is due to Bucy.[6]

Overview[edit]

Assume the state of the system evolves according to

dx = f(x,t) \, dt + \sigma dw

and a noisy measurement of the system state is available:

dz = h(x,t) \, dt + \eta dv

where w, v are independent Wiener processes. Then the conditional probability density p(xt) of the state at time t is given by the Kushner equation:

dp(x,t) = L[p(x,t)] dt + p(x,t) [h(x,t)-E_t h(x,t) ]^T \eta^{-\top}\eta^{-1} [dz-E_t h(x,t) dt].

where L p = -\sum \frac{\partial (f_i p)}{\partial x_i} + \frac{1}{2} \sum (\sigma \sigma^\top)_{i,j} \frac{\partial^2 p}{\partial x_i \partial x_j} is the Kolmogorov Forward operator and dp(x,t) = p(x,t + dt) - p(x,t) is the variation of the conditional probability.

The term dz-E_t h(x,t) dt is the innovation i.e. the difference between the measurement and its expected value.


Kalman-Bucy filter[edit]

One can simply use the Kushner equation to derive the Kalman-Bucy filter for a linear diffusion process. Suppose we have  f(x,t) = a x and  h(x,t) = c x . The Kushner equation will be given by


dp(x,t) = L[p(x,t)] dt + p(x,t) [c x- c \mu(t)]^T \eta^{-\top}\eta^{-1} [dz-c \mu(t) dt],

where  \mu(t) is the mean of the conditional probability at time  t. Multiplying by  x and integrating over it, we obtain the variation of the mean


d\mu(t) = a \mu(t) dt + \Sigma(t) c^\top \eta^{-\top}\eta^{-1} \left(dz - c\mu(t) dt\right).

Likewise, the variation of the variance \Sigma(t) is given by


\frac{d\Sigma(t)}{dt} = a\Sigma(t) + \Sigma(t) a^\top + \sigma^\top \sigma-\Sigma(t) c^\top\eta^{-\top} \eta^{-1} c \Sigma(t).

The conditional probability is then given at every instant by a normal distribution \mathcal{N}(\mu(t),\Sigma(t)).

References[edit]

  1. ^ Kushner H.J. (1964) On the differential equations satisfied by conditional probability densities of Markov processes, with applications.. J. SIAM Control Ser. A, 2(1), pp. 106-119.
  2. ^ Stratonovich, R.L. (1959). Optimum nonlinear systems which bring about a separation of a signal with constant parameters from noise. Radiofizika, 2:6, pp. 892–901.
  3. ^ Stratonovich, R.L. (1959). On the theory of optimal non-linear filtering of random functions. Theory of Probability and its Applications, 4, pp. 223–225.
  4. ^ Stratonovich, R.L. (1960) Application of the Markov processes theory to optimal filtering. Radio Engineering and Electronic Physics, 5:11, pp. 1–19.
  5. ^ Stratonovich, R.L. (1960). Conditional Markov Processes. Theory of Probability and its Applications, 5, pp. 156–178.
  6. ^ Bucy, R. S. (1965) Nonlinear filtering theory. IEEE Transactions on Automatic Control, 10, pp. 198–198. [1]

See also[edit]