# Unscented transform

Jump to: navigation, search

The Unscented Transform (or UT) is a mathematical function used to estimate the result of applying a given nonlinear transformation to a probability distribution that is characterized only in terms of a finite set of statistics. The most common use of the Unscented Transform is in the nonlinear projection of mean and covariance estimates in the context of nonlinear extensions of the Kalman filter. In an interview,[1] its creator Jeffrey Uhlmann explained that he came up with the name after noticing unscented deodorant on a coworker's desk.

## Background

Many filtering and control methods represent estimates of the state of a system in the form of a mean vector and an associated error covariance matrix. As an example, the estimated 2-dimensional position of an object of interest might be represented by a mean position vector, $[x, y]$, with an uncertainty given in the form of a 2x2 covariance matrix giving the variance in $x$, the variance in $y$, and the cross covariance between the two. A covariance that is zero implies that there is no uncertainty or error and that the position of the object is exactly what is specified by the mean vector.

The mean and covariance representation only gives the first two moments of an underlying, but otherwise unknown, probability distribution. In the case of a moving object, the unknown probability distribution might represent the uncertainty of the object's position at a given time. The mean and covariance representation of uncertainty is mathematically convenient because any linear transformation $T$ can be applied to a mean vector $m$ and covariance matrix $M$ as $Tm$ and $TMT^T$. This linearity property does not hold for moments beyond the first central moment (the mean) and the second central moment (the covariance), so it is not generally possible to determine the mean and covariance resulting from a nonlinear transformation because the result depends on all the moments, and only the first two are given.

Although the covariance matrix is often treated as being the expected squared error associated with the mean, in practice the matrix is maintained as an upper bound on the actual squared error. Specifically, a mean and covariance estimate $(m,M)$ is conservatively maintained so that the covariance matrix $M$ is greater than or equal to the actual squared error associated with $m$. Mathematically this means that the result of subtracting the expected squared error (which is not usually known) from $M$ is a semidefinite or positive-definite matrix. The reason for maintaining a conservative covariance estimate is that most filtering and control algorithms will tend to diverge (fail) if the covariance is underestimated. This is because a spuriously small covariance implies less uncertainty and leads the filter to place more weight (confidence) than is justified in the accuracy of the mean.

Returning to the example above, when the covariance is zero it is trivial to determine the location of the object after it moves according to an arbitrary nonlinear function $f(x,y)$: just apply the function to the mean vector. When the covariance is not zero the transformed mean will not generally be equal to $f(x,y)$ and it is not even possible to determine the mean of the transformed probability distribution from only its prior mean and covariance. Given this indeterminacy, the nonlinearly transformed mean and covariance can only be approximated. The earliest such approximation was to linearize the nonlinear function and apply the resulting Jacobian matrix to the given mean and covariance. This is the basis of the Extended Kalman Filter (EKF), and although it was known to yield poor results in many circumstances, there was no practical alternative for many decades.

## Motivation for the Unscented Transform

In 1994 Jeffrey Uhlmann noted that the EKF takes a nonlinear function and partial distribution information (in the form of a mean and covariance estimate) of the state of a system but applies an approximation to the known function rather than to the imprecisely-known probability distribution. He suggested that a better approach would be to use the exact nonlinear function applied to an approximating probability distribution. The motivation for this approach is given in his doctoral dissertation, where the term Unscented Transform was first defined:[2]

Consider the following intuition: With a fixed number of parameters it should be easier to approximate a given distribution than it is to approximate an arbitrary nonlinear function/transformation. Following this intuition, the goal is to find a parameterization that captures the mean and covariance information while at the same time permitting the direct propagation of the information through an arbitrary set of nonlinear equations. This can be accomplished by generating a discrete distribution having the same first and second (and possibly higher) moments, where each point in the discrete approximation can be directly transformed. The mean and covariance of the transformed ensemble can then be computed as the estimate of the nonlinear transformation of the original distribution. More generally, the application of a given nonlinear transformation to a discrete distribution of points, computed so as to capture a set of known statistics of an unknown distribution, is referred to as an unscented transformation.

In other words, the given mean and covariance information can be exactly encoded in a set of points, referred to as sigma points, which if treated as elements of a discrete probability distribution has mean and covariance equal to the given mean and covariance. This distribution can be propagated exactly by applying the nonlinear function to each point. The mean and covariance of the transformed set of points then represents the desired transformed estimate. The principal advantage of the approach is that the nonlinear function is fully exploited, as opposed to the EKF which replaces it with a linear one. Eliminating the need for linearization also provides advantages independent of any improvement in estimation quality. One immediate advantage is that the UT can be applied with any given function whereas linearization may not be possible for functions that are not differentiable. A practical advantage is that the UT can be easier to implement because it avoids the need to derive and implement a linearizing Jacobian matrix.

## Computing the Unscented Transform

Uhlmann noted that $n+1$ sigma points are necessary and sufficient in the general case to define a discrete distribution having a given mean and covariance in $n$ dimensions.[2] Consider the following simplex of points in two dimensions:

$s_1 = \left[0, \sqrt{2}\right]^T, \quad s_2 = \left[-\sqrt{3\over 2}, -\sqrt{1\over 2}\right]^T, \quad s_3 = \left[\sqrt{3\over 2}, -\sqrt{1\over 2}\right]^T$

It can be verified that the above set of points has mean $s=0$ and covariance $S=I$ (the identity matrix). Given any 2-dimensional mean and covariance, $(x, X)$, the desired sigma points can be obtained by multiplying each point by the matrix square root of $X$ and adding $x$. A similar canonical set of sigma points can be generated in any number of dimensions $n$ by taking the zero vector and the points comprising the rows of the identity matrix, computing the mean of the set of points, subtracting the mean from each point so that the resulting set has a mean of zero, then compute the covariance of the zero-mean set of points and apply its inverse to each point so that the covariance of the set will be equal to the identity.

Uhlmann showed that it is possible to conveniently generate a symmetric set of $2n$ sigma points from the columns of $\pm\sqrt{nX}$, where $X$ is the given covariance matrix, without having to compute a matrix inverse. It is computationally efficient and, because the points form a symmetric distribution, captures the third central moment (the skew) whenever the underlying distribution of the state estimate is known or can be assumed to be symmetric.[2] He also showed that weights, including negative weights, can be used to affect the statistics of the set. He and Simon Julier published several papers showing that the use of the Unscented transformation in a Kalman filter, which is referred to as the Unscented Kalman Filter (UKF), provides significant performance improvements over the EKF in a variety of applications.[3][4][5] Julier also developed and examined techniques for generating sigma points to capture the third moment (the skew) of an arbitrary distribution and the fourth moment (the kurtosis) of a symmetric distribution.[4][6]

It should be noted that Julier and Uhlmann published papers using a particular parameterized form of the Unscented Transform in the context of the UKF which used negative weights to capture assumed distribution information.[3][5] That form of the UT is susceptible to a variety of numerical errors that the original formulations (above) do not suffer. Julier has subsequently described parameterized forms which do not use negative weights and also are not subject to those issues.[7]

## Example

The Unscented Transform is defined for the application of a given function to any partial characterization of an otherwise unknown distribution, but its most common use is for the case in which only the mean and covariance is given. A common example is the conversion from one coordinate system to another, such as from a Cartesian coordinate frame to polar coordinates.[4]

Suppose a 2-dimensional mean and covariance estimate, $(m, M)$, is given in Cartesian coordinates with:

$m = [12.3, 7.6]^T, \quad M = \begin{bmatrix}1.44 & 0 \\0 & 2.89\end{bmatrix}$

and the transformation function to polar coordinates, $f(x, y) \rightarrow [r, \theta]$, is:

$r = \sqrt{(x^2+y^2)}, \quad \theta = \arctan(y/x)$

Multiplying each of the canonical simplex sigma points (given above) by $M^{1/2}=\begin{bmatrix}1.2 & 0 \\0 & 1.7\end{bmatrix}$ and adding the mean, $m$, gives:

$m_1 = [0, 2.40] + [12.3, 7.6] = [12.3, 10.0]$
$m_2 = [-1.47, -1.20] + [12.3, 7.6] = [10.8, 6.40]$
$m_3 = [1.47, -1.20] + [12.3, 7.6] = [13.8, 6.40]$

Applying the transformation function $f()$ to each of the above points gives:

${m^+}_1 = f(12.3, 10.0) = [15.85, 0.68]$
${m^+}_2 = f(10.8, 6.40) = [12.58, 0.53]$
${m^+}_3 = f(13.8, 6.40) = [15.18, 0.44]$

The mean of these three transformed points, $m_{UT}=\frac{1}{3}\Sigma^3_{i=1}{m^+}_i$, is the UT estimate of the mean in polar coordinates:

$m_{UT} = [14.539, 0.551]$

The UT estimate of the covariance is:

$M_{UT}=\frac{1}{3}\Sigma^3_{i=1}({m^+}_i-m_{UT})^2$

where the each squared term in the sum is a vector outer product. This gives:

$M_{UT} = \begin{bmatrix}2.00 & 0.0443 \\0.0443 & 0.0104\end{bmatrix}$

This can be compared to the linearized mean and covariance:

$m_\mbox{linear} = f(12.3, 7.6) = [14.46,0.554]^T$
$M_\mbox{linear} = \nabla_f M \nabla_f^T = \begin{bmatrix}1.927 & 0.047 \\0.047 & 0.011\end{bmatrix}$

The absolute difference between the UT and linearized estimates in this case is relatively small, but in filtering applications the cumulative effect of small errors can lead to unrecoverable divergence of the estimate. The effect of the errors are exacerbated when the covariance is underestimated because this causes the filter to be overconfident in the accuracy of the mean. In the above example it can be seen that the linearized covariance estimate is smaller than that of the UT estimate, suggesting that linearization has likely produced an underestimate of the actual error in its mean.

In this example there is no way to determine the absolute accuracy of the UT and linearized estimates without ground truth in the form of the actual probability distribution associated with the original estimate and the mean and covariance of that distribution after application of the nonlinear transformation (e.g., as determined analytically or through numerical integration). Such analyses have been performed for coordinate transformations under the assumption of Gaussianity for the underlying distributions, and the UT estimates tend to be significantly more accurate than those obtained from linearization.[5][8]

Empirical analysis has shown that the use of the minimal simplex set of $n+1$ sigma points is significantly less accurate than the use of the symmetric set of $2n$ points when the underlying distribution is Gaussian.[8] This suggests that the use of the simplex set in the above example would not be the best choice if the underlying distribution associated with $(m,M)$ is symmetric. Even if the underlying distribution is not symmetric, the simplex set is still likely to be less accurate than the symmetric set because the asymmetry of the simplex set is not matched to the asymmetry of the actual distribution.

Returning to the example, the minimal symmetric set of sigma points can be obtained from the covariance matrix $M=\begin{bmatrix}1.44 & 0 \\0 & 2.89\end{bmatrix}$ simply as the mean vector, $m=[12.3, 7.6]$ plus and minus the columns of $(2M)^{1/2}=\sqrt{2}*\begin{bmatrix}1.2 & 0 \\0 & 1.7\end{bmatrix}=\begin{bmatrix}1.697 & 0 \\0 &2.404\end{bmatrix}$:

$m_1 = [12.3, 7.6] + [1.697, 0] = [13.997, 7.6]$
$m_2 = [12.3, 7.6] - [1.697, 0] = [10.603, 7.6]$
$m_3 = [12.3, 7.6] + [0, 2.404] = [12.3, 10.004]$
$m_4 = [12.3, 7.6] - [0, 2.404] = [12.3, 5.196]$

This construction guarantees that the mean and covariance of the above four sigma points is $(m,M)$, which is directly verifiable. Applying the nonlinear function $f()$ to each of the sigma points gives:

${m^+}_1 = [15.927, 0.497]$
${m^+}_2 = [13.045, 0.622]$
${m^+}_3 = [15.854, 0.683]$
${m^+}_4 = [13.352, 0.400]$

The mean of these four transformed sigma points, $m_{UT}=\frac{1}{4}\Sigma^4_{i=1}{m'}_i$, is the UT estimate of the mean in polar coordinates:

$m_{UT} = [14.545, 0.550]$

The UT estimate of the covariance is:

$M_{UT}=\frac{1}{4}\Sigma^4_{i=1}({m^+}_i-m_{UT})^2$

where the each squared term in the sum is a vector outer product. This gives:

$M_{UT} = \begin{bmatrix}1.823 & 0.043 \\0.043 & 0.012\end{bmatrix}$

The difference between the UT and linearized mean estimates gives a measure of the effect of the nonlinearity of the transformation. When the transformation is linear, for instance, the UT and linearized estimates will be identical. This motivates the use of the square of this difference to be added to the UT covariance to guard against underestimating of the actual error in the mean. This approach does not improve the accuracy of the mean but can significantly improve the accuracy of a filter over time by reducing the likelihood that the covariance is underestimated.[2]

## Optimality of the Unscented Transform

Uhlmann noted that given only the mean and covariance of an otherwise unknown probability distribution, the transformation problem is ill-defined because there are an infinite number of possible underlying distributions with the same first two moments. Without any a priori information or assumptions about the characteristics of the underlying distribution, any choice of distribution used to compute the transformed mean and covariance is as reasonable as any other. In other words, there is no choice of distribution with a given mean and covariance that is superior to that provided by the set of sigma points, therefore the Unscented Transform is trivially optimal.

This general statement of optimality is of course useless for making any quantitative statements about the performance of the UT, e.g., compared to linearization; consequently he, Julier, and others have performed analyses under various assumptions about the characteristics of the distribution and/or the form of the nonlinear transformation function. For example, if the function is differentiable, which is essential for linearization, these analyses validate the expected and empirically-corroborated superiority of the Unscented Transform.[5][8]

## Applications

The Unscented Transform, especially as part of the UKF, has largely replaced the EKF in many nonlinear filtering and control applications, including for underwater,[9] ground and air navigation,[10] and spacecraft.[11]

## References

1. ^ http://www.ieeeghn.org/wiki/index.php/First-Hand:The_Unscented_Transform
2. ^ a b c d Uhlmann, Jeffrey (1995). Dynamic Map Building and Localization: New Theoretical Foundations (Ph.D. thesis). University of Oxford.
3. ^ a b Julier, S.; J. Uhlmann (1997). "New Extension of the Kalman Filter to Nonlinear Systems". Proceedings of the 1997 SPIE Conference on Signal Processing, Sensor Fusion, and Target Recognition 3068.
4. ^ a b c Julier, S.; J. Uhlmann (1997). "Consistent Debiased Method for Converting Between Polar and Cartesian Coordinate Systems". Proceedings of the 1997 SPIE Conference on Acquisition, Tracking, and Pointing 3086. SPIE.
5. ^ a b c d Julier, Simon; Uhlmann, Jeffrey (2000). "A New Method for the Nonlinear Transformation of Means and Covariances in Nonlinear Filters". IEEE Trans. On Automatic Control.
6. ^ Julier, Simon (1998). "A Skewed Approach to Filtering". The Proceedings of the 12th Intl. Symp. On Aerospace/Defense Sensing, Simulation and Controls 3373. SPIE.
7. ^ Julier, Simon (2002). "The Scaled Unscented Transformation". Proceedings of the American Control Conference 6. IEEE.
8. ^ a b c Zhang, W.; M. Liu and Z. Zhao (2009). "Accuracy Analysis of Unscented Transformation of Several Sampling Strategies". Proc. of the 10th Intl. Conf. on Software Engineering, Artificial Intelligence, Networking and Parallel/Distributed Computing. ACIS.
9. ^ Wu, L.; J. Ma and J. Tian (2010). "Self-Adaptive Unscented Kalman Filtering for Underwater Gravity Aided Navigation". Proc. of IEEE/ION Plans.
10. ^ El-Sheimy, N; Shin, EH; Niu, X (2006). "Kalman Filter Face-Off: Extended vs. Unscented Kalman Filters for Integrated GPS and MEMS Inertial". Inside GNSS: Engineering Solutions for the Global Navigation Satellite System Community 1 (2).
11. ^ Crassidis, J.; Markley, F. (2003). "Unscented Filtering for Spacecraft Attitude Estimation". AIAA Journal of Guidance, Control, and Dynamics 26.