In mathematics, the matrix exponential is a matrix function on square matrices analogous to the ordinary exponential function. Abstractly, the matrix exponential gives the connection between a matrix Lie algebra and the corresponding Lie group.
Let X be an n×n real or complex matrix. The exponential of X, denoted by eX or exp(X), is the n×n matrix given by the power series
![{\displaystyle e^{X}=\sum _{k=0}^{\infty }{X^{k} \over k!}.}](https://wikimedia.org/api/rest_v1/media/math/render/svg/3aa0ea0287dd01812679ca9786e1aed0139b8ab5)
The above series always converges, so the exponential of X is well-defined. Note that if X is a 1×1 matrix the matrix exponential of X is a 1×1 matrix consisting of the ordinary exponential of the single element of X.
Properties
Let X and Y be n×n complex matrices and let a and b be arbitrary complex numbers. We denote the n×n identity matrix by
and the zero matrix by 0. The matrix exponential satisfies the following properties:
![{\displaystyle e^{0}=I.\,}](https://wikimedia.org/api/rest_v1/media/math/render/svg/9b0793d591f175f8ba9fe88084cd22be85c411b4)
![{\displaystyle e^{aX}e^{bX}=e^{(a+b)X}.\,}](https://wikimedia.org/api/rest_v1/media/math/render/svg/0ffc058fd9ee851df1101dd70c5ed9887e642675)
![{\displaystyle e^{X}e^{-X}=I.\,}](https://wikimedia.org/api/rest_v1/media/math/render/svg/21a444df4e85a7ad82cae3070a75544b8c682d8c)
- If
then ![{\displaystyle e^{A}e^{B}=e^{A+B}.\,}](https://wikimedia.org/api/rest_v1/media/math/render/svg/b46b2766bef13950e6d41ed76410a6fbb72b44ff)
- If
is invertible then ![{\displaystyle e^{YXY^{-1}}=Ye^{X}Y^{-1}.}](https://wikimedia.org/api/rest_v1/media/math/render/svg/306dd35cdc74299df998f5cc95ac3bb2e2972e93)
, where tr(X) is the trace of X.
- exp(XT) = (exp X)T, where XT denotes the transpose of X. It follows that if X is symmetric then eX is also symmetric, and that if X is skew-symmetric then eX is orthogonal.
- exp(X*) = (exp X)*, where X* denotes the conjugate transpose of X. It follows that if X is Hermitian then eX is also Hermitian, and that if X is skew-Hermitian then eX is unitary.
Linear differential equations
One of the reasons for the importance of the matrix exponential is that it can be used to solve systems of linear ordinary differential equations. Indeed, it follows from equation (1) below that the solution of
![{\displaystyle {\frac {d}{dt}}y(t)=Ay(t),\quad y(0)=y_{0},}](https://wikimedia.org/api/rest_v1/media/math/render/svg/e5aece55f011795ab4179383e7f5bd5d06ac6841)
where A is a matrix, is given by
![{\displaystyle y(t)=e^{At}y_{0}.\,}](https://wikimedia.org/api/rest_v1/media/math/render/svg/0ebab5d216a41242ee68745367921995d3c63f4e)
The matrix exponential can also be used to solve the inhomogeneous equation
![{\displaystyle {\frac {d}{dt}}y(t)=Ay(t)+z(t),\quad y(0)=y_{0}.}](https://wikimedia.org/api/rest_v1/media/math/render/svg/c2e65d9e8b388b0b0da52a90f3adfc75add82b35)
See the section on applications below for examples.
There is no closed-form solution for differential equations of the form
![{\displaystyle {\frac {d}{dt}}y(t)=A(t)\,y(t),\quad y(0)=y_{0},}](https://wikimedia.org/api/rest_v1/media/math/render/svg/88f0a9208671905c73c384c2f4d19bdc0cb17257)
where A is not constant, but the Magnus series gives the solution as an infinite sum.
The exponential of sums
We know that the exponential function satisfies
for any real numbers (scalars) x and y. The same goes for commuting matrices: If the matrices X and Y commute (meaning that XY = YX), then
![{\displaystyle e^{X+Y}=e^{X}e^{Y}.\,}](https://wikimedia.org/api/rest_v1/media/math/render/svg/81e55c9227fd61b2fc716c91b14eab16a76625b8)
However, if they do not commute, then the above equality does not necessarily hold. In that case, we can use the Baker-Campbell-Hausdorff formula to compute
.
The exponential map
Note that the exponential of a matrix is always a non-singular matrix. The inverse of eX is given by e−X. This is analogous to the fact that the exponential of a complex number is always nonzero. The matrix exponential then gives us a map
![{\displaystyle \exp \colon M_{n}(\mathbb {C} )\to {\mbox{GL}}(n,\mathbb {C} )}](https://wikimedia.org/api/rest_v1/media/math/render/svg/c0ec994e7437c7b5a857927ccc2e7ffc9457d1e0)
from the space of all n×n matrices to the general linear group, i.e. the group of all non-singular matrices. In fact, this map is surjective which means that every non-singular matrix can be written as the exponential of some other matrix (for this, it is essential to consider the field C of complex numbers and not R). The matrix logarithm gives an inverse to this map.
For any two matrices X and Y, we have
![{\displaystyle \|e^{X+Y}-e^{X}\|\leq \|Y\|e^{\|X\|}e^{\|Y\|},}](https://wikimedia.org/api/rest_v1/media/math/render/svg/78a1fb5fd5dfe94763704fff01bc554d1d351430)
where || · || denotes an arbitrary matrix norm. It follows that the exponential map is continuous and Lipschitz continuous on compact subsets of Mn(C).
The map
![{\displaystyle t\mapsto e^{tX},\qquad t\in \mathbb {R} }](https://wikimedia.org/api/rest_v1/media/math/render/svg/79c8cfd3a6dbdcf5b4fefb10c39c679dd72db266)
defines a smooth curve in the general linear group which passes through the identity element at t = 0. In fact, this gives a one-parameter subgroup of the general linear group since
![{\displaystyle e^{tX}e^{sX}=e^{(t+s)X}.\,}](https://wikimedia.org/api/rest_v1/media/math/render/svg/d6eb043cb73b775abc0022d6298cea1f6e3e62e9)
The derivative of this curve (or tangent vector) at a point t is given by
![{\displaystyle {\frac {d}{dt}}e^{tX}=Xe^{tX}.\qquad (1)}](https://wikimedia.org/api/rest_v1/media/math/render/svg/68000dc615fa949f79e6b4dfb602f25508892dfd)
The derivative at t = 0 is just the matrix X, which is to say that X generates this one-parameter subgroup.
More generally,
![{\displaystyle {\frac {d}{dt}}e^{X(t)}=\int _{0}^{1}e^{(1-\alpha )X(t)}{\frac {dX(t)}{dt}}e^{\alpha X(t)}\,d\alpha .}](https://wikimedia.org/api/rest_v1/media/math/render/svg/5f33e0349bd73d55b72acc391b2cec72032993c4)
Computing the matrix exponential
Finding reliable and accurate methods to compute the matrix exponential is difficult, and this is still a topic of considerable current research in mathematics and numerical analysis. Some methods are listed below.
Diagonalizable case
If a matrix is diagonal:
![{\displaystyle A={\begin{bmatrix}a_{1}&0&\ldots &0\\0&a_{2}&\ldots &0\\\vdots &\vdots &\ddots &\vdots \\0&0&\ldots &a_{n}\end{bmatrix}},}](https://wikimedia.org/api/rest_v1/media/math/render/svg/7d43bd52fe793eee3c5d98f60e5d1f5f8b1a5fb8)
then its exponential can be obtained by just exponentiating every entry on the main diagonal:
![{\displaystyle e^{A}={\begin{bmatrix}e^{a_{1}}&0&\ldots &0\\0&e^{a_{2}}&\ldots &0\\\vdots &\vdots &\ddots &\vdots \\0&0&\ldots &e^{a_{n}}\end{bmatrix}}.}](https://wikimedia.org/api/rest_v1/media/math/render/svg/0fc981f695865a810ecb811aa83e231f52f01488)
This also allows one to exponentiate diagonalizable matrices. If A = UDU−1 and D is diagonal, then eA = UeDU−1. Application of Sylvester's formula yields the same result.
Nilpotent case
A matrix N is nilpotent if Nq = 0 for some integer q. In this case, the matrix exponential eN can be computed directly from the series expansion, as the series terminates after a finite number of terms:
![{\displaystyle e^{N}=I+N+{\frac {1}{2}}N^{2}+{\frac {1}{6}}N^{3}+\cdots +{\frac {1}{(q-1)!}}N^{q-1}.}](https://wikimedia.org/api/rest_v1/media/math/render/svg/14700f7490cbc68f87e50150fd1b95e7819c4dfc)
General case
An arbitrary matrix X (over an arbitrary field) can be expressed uniquely as sum
![{\displaystyle X=A+N\,}](https://wikimedia.org/api/rest_v1/media/math/render/svg/f6c484f44729510798f166ff8b1d7afc35748a5d)
where
- A is diagonalizable
- N is nilpotent
- A commutes with N (i.e. AN = NA)
This means we can compute the exponential of X by reducing to the previous two cases:
![{\displaystyle e^{X}=e^{A+N}=e^{A}e^{N}.\,}](https://wikimedia.org/api/rest_v1/media/math/render/svg/24ca1b3e97921e1aa0237d2d4561aaa8fc2ce327)
Note that we need the commutativity of A and N for the last step to work.
Another (closely related) method if the field is algebraically closed is to work with the Jordan form of X. Suppose that X = PJP −1 where J is the Jordan form of X. Then
![{\displaystyle e^{X}=Pe^{J}P^{-1}.\,}](https://wikimedia.org/api/rest_v1/media/math/render/svg/59b15084eefab1055e349f80b1fefc1a33cb4596)
Also, since
![{\displaystyle J=J_{a_{1}}(\lambda _{1})\oplus J_{a_{2}}(\lambda _{2})\oplus \cdots \oplus J_{a_{n}}(\lambda _{n}),}](https://wikimedia.org/api/rest_v1/media/math/render/svg/c7ab8d06cd720746d25050b025354f0ffeab3dc2)
![{\displaystyle {\begin{aligned}e^{J}&{}=\exp {\big (}J_{a_{1}}(\lambda _{1})\oplus J_{a_{2}}(\lambda _{2})\oplus \cdots \oplus J_{a_{n}}(\lambda _{n}){\big )}\\&{}=\exp {\big (}J_{a_{1}}(\lambda _{1}){\big )}\oplus \exp {\big (}J_{a_{2}}(\lambda _{2}){\big )}\oplus \cdots \oplus \exp {\big (}J_{a_{k}}(\lambda _{k}){\big )}.\end{aligned}}}](https://wikimedia.org/api/rest_v1/media/math/render/svg/3440644c09e276ec39d481d6124c96269b2b01fb)
Therefore, we need only know how to compute the matrix exponential of a Jordan block. But each Jordan block is of the form
![{\displaystyle J_{a}(\lambda )=\lambda I+N\,}](https://wikimedia.org/api/rest_v1/media/math/render/svg/8e69875083d7679a23d73cefd30d6063780f0c5f)
where N is a special nilpotent matrix. The matrix exponential of this block is given by
![{\displaystyle e^{\lambda I+N}=e^{\lambda }e^{N}.\,}](https://wikimedia.org/api/rest_v1/media/math/render/svg/6feac4b6dc45b3fe471dfacf9483c091568f77da)
Calculations
Suppose that we want to compute the exponential of
![{\displaystyle B={\begin{bmatrix}21&17&6\\-5&-1&-6\\4&4&16\end{bmatrix}}.}](https://wikimedia.org/api/rest_v1/media/math/render/svg/9d3277007480f8371776c095ab3befa94b7a53d2)
Its Jordan form is
![{\displaystyle J=P^{-1}BP={\begin{bmatrix}4&0&0\\0&16&1\\0&0&16\end{bmatrix}},}](https://wikimedia.org/api/rest_v1/media/math/render/svg/580ce84882ce214476b055d635cd9aedd5074e63)
where the matrix P is given by
![{\displaystyle P={\begin{bmatrix}-{\frac {1}{4}}&2&{\frac {5}{4}}\\{\frac {1}{4}}&-2&-{\frac {1}{4}}\\0&4&0\end{bmatrix}}.}](https://wikimedia.org/api/rest_v1/media/math/render/svg/4e371494c6a51b6d5a5ef8f7744cf018ce2afdb3)
Let us first calculate exp(J). We have
![{\displaystyle J=J_{1}(4)\oplus J_{2}(16)}](https://wikimedia.org/api/rest_v1/media/math/render/svg/88f45865d147222d8b16ef38760e91c30aded3dc)
The exponential of a 1×1 matrix is just the exponential of the one entry of the matrix, so exp(J1(4)) = [e4]. The exponential of
can be calculated by the formula exp(λ
+N) = eλ exp(N) mentioned above; this yields
![{\displaystyle \exp \left({\begin{bmatrix}16&1\\0&16\end{bmatrix}}\right)=e^{16}\exp \left({\begin{bmatrix}0&1\\0&0\end{bmatrix}}\right)=e^{16}\left({\begin{bmatrix}1&0\\0&1\end{bmatrix}}+{\begin{bmatrix}0&1\\0&0\end{bmatrix}}+{1 \over 2!}{\begin{bmatrix}0&0\\0&0\end{bmatrix}}+\cdots \right)={\begin{bmatrix}e^{16}&e^{16}\\0&e^{16}\end{bmatrix}}.}](https://wikimedia.org/api/rest_v1/media/math/render/svg/46289f37b9c99cb679ecd21fa36f2b88fe761932)
Therefore, the exponential of the original matrix B is
![{\displaystyle \exp(B)=P\exp(J)P^{-1}=P{\begin{bmatrix}e^{4}&0&0\\0&e^{16}&e^{16}\\0&0&e^{16}\end{bmatrix}}P^{-1}={1 \over 4}{\begin{bmatrix}13e^{16}-e^{4}&13e^{16}-5e^{4}&2e^{16}-2e^{4}\\-9e^{16}+e^{4}&-9e^{16}+5e^{4}&-2e^{16}+2e^{4}\\16e^{16}&16e^{16}&4e^{16}\end{bmatrix}}.}](https://wikimedia.org/api/rest_v1/media/math/render/svg/ed778b8c7c4ef619b23f766dc86c48a157fc401a)
Applications
Linear differential equations
The matrix exponential has applications to systems of linear differential equations. Recall from earlier in this article that a differential equation of the form
- y′ = Cy
has solution eCxy(0). If we consider the vector
![{\displaystyle \mathbf {y} (x)={\begin{pmatrix}y_{1}(x)\\\vdots \\y_{n}(x)\end{pmatrix}}}](https://wikimedia.org/api/rest_v1/media/math/render/svg/7f57dc8fba7859613058676625761748a5a5d536)
we can express a system of coupled linear differential equations as
![{\displaystyle \mathbf {y} '(x)=A\mathbf {y} (x)+\mathbf {b} }](https://wikimedia.org/api/rest_v1/media/math/render/svg/581f80f31a6adf3873b55a13bea8d53a3bd12a75)
If we make an ansatz and use an integrating factor of e−Ax and multiply throughout, we obtain
![{\displaystyle e^{-Ax}\mathbf {y} '(x)-e^{-Ax}A\mathbf {y} =e^{-Ax}\mathbf {b} }](https://wikimedia.org/api/rest_v1/media/math/render/svg/9f3159c752842e7b962193028263d543edd9838d)
![{\displaystyle D(e^{-Ax}\mathbf {y} )=e^{-Ax}\mathbf {b} }](https://wikimedia.org/api/rest_v1/media/math/render/svg/eddec327430967dbfbce77020a753b611fe55994)
If we can calculate eAx, then we can obtain the solution to the system.
Example (homogeneous)
Say we have the system
![{\displaystyle {\begin{matrix}x'&=&2x&-y&+z\\y'&=&&3y&-1z\\z'&=&2x&+y&+3z\end{matrix}}}](https://wikimedia.org/api/rest_v1/media/math/render/svg/e6d4792379d8eef0156649a4d6d008c304794807)
We have the associated matrix
![{\displaystyle M={\begin{bmatrix}2&-1&1\\0&3&-1\\2&1&3\end{bmatrix}}}](https://wikimedia.org/api/rest_v1/media/math/render/svg/8e69a49c6ba7d4a025867a570b62b3d826b4cc3f)
In the example above, we have calculated the matrix exponential
![{\displaystyle e^{tM}={\begin{bmatrix}2e^{t}-2te^{2t}&-2te^{2t}&0\\-2e^{t}+2(t+1)e^{2t}&2(t+1)e^{2t}&0\\2te^{2t}&2te^{2t}&2e^{t}\end{bmatrix}}}](https://wikimedia.org/api/rest_v1/media/math/render/svg/9adacee2be390d8c415b59401f7651aeb47efbd9)
so the general solution of the system is
![{\displaystyle {\begin{bmatrix}x\\y\\z\end{bmatrix}}=C_{1}{\begin{bmatrix}2e^{t}-2te^{2t}\\-2e^{t}+2(t+1)e^{2t}\\2te^{2t}\end{bmatrix}}+C_{2}{\begin{bmatrix}-2te^{2t}\\2(t+1)e^{2t}\\2te^{2t}\end{bmatrix}}+C_{3}{\begin{bmatrix}0\\0\\2e^{t}\end{bmatrix}}}](https://wikimedia.org/api/rest_v1/media/math/render/svg/6d182e7c15bf9145c255efacdc40ec6585c575d2)
that is,
![{\displaystyle {\begin{aligned}x&=C_{1}(2e^{t}-2te^{2t})+C_{2}(-2te^{2t})\\y&=C_{1}(-2e^{t}+2(t+1)e^{2t})+C_{2}(2(t+1)e^{2t})\\z&=(C_{1}+C_{2})(2te^{2t})+2C_{3}e^{t}\end{aligned}}}](https://wikimedia.org/api/rest_v1/media/math/render/svg/e55753bf089272f5c80ceddb58fd76311c7748b6)
Inhomogeneous case - variation of parameters
For the inhomogeneous case, we can use integrating factors (a method akin to variation of parameters). We seek a particular solution of the form yp(t) = exp(tA)z(t) :
![{\displaystyle \mathbf {y} _{p}'=(e^{tA})'\mathbf {z} (t)+e^{tA}\mathbf {z} '(t)}](https://wikimedia.org/api/rest_v1/media/math/render/svg/95719fe538354257c2b35c2db045a19e6f8b8d00)
![{\displaystyle =Ae^{tA}\mathbf {z} (t)+e^{tA}\mathbf {z} '(t)}](https://wikimedia.org/api/rest_v1/media/math/render/svg/29ebad2eb67e3a4a1cdc9612483f475a0032d754)
![{\displaystyle =A\mathbf {y} _{p}(t)+e^{tA}\mathbf {z} '(t)}](https://wikimedia.org/api/rest_v1/media/math/render/svg/cf16e8fb7d12738d2bcfd578c2bd079897792cda)
For yp to be a solution:
![{\displaystyle e^{tA}\mathbf {z} '(t)=\mathbf {b} (t)}](https://wikimedia.org/api/rest_v1/media/math/render/svg/ce73caf9d7f5ae8b8c57cb0912cce2b59c89d61a)
![{\displaystyle \mathbf {z} '(t)=(e^{tA})^{-1}\mathbf {b} (t)}](https://wikimedia.org/api/rest_v1/media/math/render/svg/1c95eb11e8fd08263e98afdd844635e0f03cc188)
![{\displaystyle \mathbf {z} (t)=\int _{0}^{t}e^{-uA}\mathbf {b} (u)\,du+\mathbf {c} }](https://wikimedia.org/api/rest_v1/media/math/render/svg/cd24a985f5b3e722145a3a05f2c5b9a692ea66f1)
So,
![{\displaystyle {\begin{aligned}\mathbf {y} _{p}&{}=e^{tA}\int _{0}^{t}e^{-uA}\mathbf {b} (u)\,du+e^{tA}\mathbf {c} \\&{}=\int _{0}^{t}e^{(t-u)A}\mathbf {b} (u)\,du+e^{tA}\mathbf {c} \end{aligned}}}](https://wikimedia.org/api/rest_v1/media/math/render/svg/622ff239585eff692f8cbba8fafb78ed5e7c589c)
where c is determined by the initial conditions of the problem.
Example (inhomogeneous)
Say we have the system
![{\displaystyle {\begin{matrix}x'&=&2x&-y&+z&+e^{2t}\\y'&=&&3y&-1z&\\z'&=&2x&+y&+3z&+e^{2t}\end{matrix}}}](https://wikimedia.org/api/rest_v1/media/math/render/svg/965c3f4e4aedd0b1a896fc04e4b1457730c94f9b)
So we then have
![{\displaystyle M={\begin{bmatrix}2&-1&1\\0&3&-1\\2&1&3\end{bmatrix}}}](https://wikimedia.org/api/rest_v1/media/math/render/svg/8e69a49c6ba7d4a025867a570b62b3d826b4cc3f)
and
![{\displaystyle \mathbf {b} =e^{2t}{\begin{bmatrix}1\\0\\1\end{bmatrix}}.}](https://wikimedia.org/api/rest_v1/media/math/render/svg/6df2d526c7844ba0c2db7c27b27eb9189b61de39)
From before, we have the general solution to the homogeneous equation, Since the sum of the homogeneous and particular solutions give the general solution to the inhomogeneous problem, now we only need to find the particular solution (via variation of parameters).
We have, above:
![{\displaystyle \mathbf {y} _{p}=e^{t}\int _{0}^{t}e^{(-u)A}{\begin{bmatrix}e^{2u}\\0\\e^{2u}\end{bmatrix}}\,du+e^{tA}\mathbf {c} }](https://wikimedia.org/api/rest_v1/media/math/render/svg/6e14a34514bd3147f4f352f6307ac8c357971cc0)
![{\displaystyle \mathbf {y} _{p}=e^{t}\int _{0}^{t}{\begin{bmatrix}2e^{u}-2ue^{2u}&-2ue^{2u}&0\\\\-2e^{u}+2(u+1)e^{2u}&2(u+1)e^{2u}&0\\\\2ue^{2u}&2ue^{2u}&2e^{u}\end{bmatrix}}{\begin{bmatrix}e^{2u}\\0\\e^{2u}\end{bmatrix}}\,du+e^{tA}\mathbf {c} }](https://wikimedia.org/api/rest_v1/media/math/render/svg/35dfad242e7204237b05410885ebf59aca0a50f3)
![{\displaystyle \mathbf {y} _{p}=e^{t}\int _{0}^{t}{\begin{bmatrix}e^{2u}(2e^{u}-2ue^{2u})\\\\e^{2u}(-2e^{u}+2(1+u)e^{2u})\\\\2e^{3u}+2ue^{4u}\end{bmatrix}}+e^{tA}\mathbf {c} }](https://wikimedia.org/api/rest_v1/media/math/render/svg/05c68be8bea2fed9a05ebe28d5703d7316cee268)
![{\displaystyle \mathbf {y} _{p}=e^{t}{\begin{bmatrix}-{1 \over 24}e^{3t}(3e^{t}(4t-1)-16)\\\\{1 \over 24}e^{3t}(3e^{t}(4t+4)-16)\\\\{1 \over 24}e^{3t}(3e^{t}(4t-1)-16)\end{bmatrix}}+{\begin{bmatrix}2e^{t}-2te^{2t}&-2te^{2t}&0\\\\-2e^{t}+2(t+1)e^{2t}&2(t+1)e^{2t}&0\\\\2te^{2t}&2te^{2t}&2e^{t}\end{bmatrix}}{\begin{bmatrix}c_{1}\\c_{2}\\c_{3}\end{bmatrix}}}](https://wikimedia.org/api/rest_v1/media/math/render/svg/b5a5abd6fc2c52dc5fb0cb0ce12f707d6e90eb85)
which can be further simplified to get the requisite particular solution determined through variation of parameters.
See also
References
- Roger A. Horn and Charles R. Johnson. Topics in Matrix Analysis. Cambridge University Press, 1991. ISBN 0-521-46713-6.
- C. Moler and C. Van Loan. Nineteen Dubious Ways to Compute the Exponential of a Matrix, Twenty-Five Years Later. SIAM Review 45, 3-49 (2003)
External links