Jump to content

Matrix sign function: Difference between revisions

From Wikipedia, the free encyclopedia
Content deleted Content added
→‎top: Added Wikipedia links
→‎top: minor
Line 3: Line 3:
In [[mathematics]], the '''matrix sign function''' is a [[matrix function]] on [[Square matrix|square matrices]] analogous to the complex [[sign function]].<ref name=":0">{{Cite book|last=Higham|first=Nicholas J.|url=https://www.worldcat.org/oclc/693957820|title=Functions of matrices : theory and computation|date=2008|publisher=Society for Industrial and Applied Mathematics (SIAM, 3600 Market Street, Floor 6, Philadelphia, PA 19104)|others=Society for Industrial and Applied Mathematics|isbn=978-0-89871-777-8|location=Philadelphia, Pa.|oclc=693957820|author-link1=Nicholas Higham}}</ref>
In [[mathematics]], the '''matrix sign function''' is a [[matrix function]] on [[Square matrix|square matrices]] analogous to the complex [[sign function]].<ref name=":0">{{Cite book|last=Higham|first=Nicholas J.|url=https://www.worldcat.org/oclc/693957820|title=Functions of matrices : theory and computation|date=2008|publisher=Society for Industrial and Applied Mathematics (SIAM, 3600 Market Street, Floor 6, Philadelphia, PA 19104)|others=Society for Industrial and Applied Mathematics|isbn=978-0-89871-777-8|location=Philadelphia, Pa.|oclc=693957820|author-link1=Nicholas Higham}}</ref>


It was introduced by Dr J.D. Roberts in 1971 as a tool for model reduction and for solving [[Lyapunov equation|Lyapunov]] and Algebraic [[Algebraic Riccati equation|Riccatia]] equation in a technical report of [[Cambridge University]], which was later published in a journal in 1980.<ref name=":1">{{Cite journal|last=Roberts|first=J. D.|title=Linear model reduction and solution of the algebraic Riccati equation by use of the sign function|url=https://www.tandfonline.com/doi/full/10.1080/00207178008922881|journal=International Journal of Control|language=en|volume=32|issue=4|pages=677–687|doi=10.1080/00207178008922881|issn=0020-7179}}</ref><ref name=":2">{{Cite journal|last=Denman|first=Eugene D.|last2=Beavers|first2=Alex N.|date=1976|title=The matrix sign function and computations in systems|url=https://www.sciencedirect.com/science/article/abs/pii/0096300376900205|journal=Applied Mathematics and Computation|language=en|volume=2|issue=1|pages=63–94|doi=10.1016/0096-3003(76)90020-5|issn=0096-3003}}</ref>
It was introduced by J.D. Roberts in 1971 as a tool for model reduction and for solving [[Lyapunov equation|Lyapunov]] and Algebraic [[Algebraic Riccati equation|Riccatia]] equation in a technical report of [[Cambridge University]], which was later published in a journal in 1980.<ref name=":1">{{Cite journal|last=Roberts|first=J. D.|title=Linear model reduction and solution of the algebraic Riccati equation by use of the sign function|url=https://www.tandfonline.com/doi/full/10.1080/00207178008922881|journal=International Journal of Control|language=en|volume=32|issue=4|pages=677–687|doi=10.1080/00207178008922881|issn=0020-7179}}</ref><ref name=":2">{{Cite journal|last=Denman|first=Eugene D.|last2=Beavers|first2=Alex N.|date=1976|title=The matrix sign function and computations in systems|url=https://www.sciencedirect.com/science/article/abs/pii/0096300376900205|journal=Applied Mathematics and Computation|language=en|volume=2|issue=1|pages=63–94|doi=10.1016/0096-3003(76)90020-5|issn=0096-3003}}</ref>


== Definition ==
== Definition ==

Revision as of 11:32, 3 April 2022

In mathematics, the matrix sign function is a matrix function on square matrices analogous to the complex sign function.[1]

It was introduced by J.D. Roberts in 1971 as a tool for model reduction and for solving Lyapunov and Algebraic Riccatia equation in a technical report of Cambridge University, which was later published in a journal in 1980.[2][3]

Definition

The matrix sign function is a generalization of the complex signum function

to the matrix valued analogue . Although the sign function is not analytic, the matrix function is well defined for all matrices that have no eigenvalue on the imaginary axis, see for example the Jordan-form-based definition (where the derivatives are all zero).

Properties

Theorem: Let , then .[1]

Theorem: Let , then is diagonalizable and has eigenvalues that are .[1]

Theorem: Let , then is a projector onto the invariant subspace associated with the eigenvalues in the right-half plane, and analogously for and the left-half plane.[1]

Theorem: Let , and be a Jordan decomposition such that corresponds to eigenvalues with positive real part and to eigenvalue with negative real part. Then , where and are identity matrices of sizes corresponding to and , respectively.[1]

Computational methods

The function can be computed with generic methods for matrix functions, but there are also specialized methods.

Newton iteration

The Newton iteration can be derived by observing that , which in terms of matrices can be written as , where we use the matrix square root. If we apply the Babylonian method to compute the square root of the matrix , that is, the iteration , and define the new iterate , we arrive at the iteration

,

where typically . Convergence is global, and locally it is quadratic.[1][2]

The Newton iteration uses the explicit inverse of the iterates .

Newton–Schulz iteration

To avoid the need of an explicit inverse used in the Newton iteration, the inverse can be approximated with one step of the Newton iteration for the inverse, , derived by Schulz(de) in 1933.[4] Substituting this approximation into the previous method, the new method becomes

.

Convergence is (still) quadratic, but only local (guaranteed for ).[1]

Applications

Solutions of Sylvester equations

Theorem:[2][3] Let and assume that and are stable, then the unique solution to the Sylvester equation, , is given by such that

Proof sketch: The result follows from the similarity transform

since

due to the stability of and .

The theorem is, naturally, also applicable to the Lyapunov equation. However, due to the structure the Newton iteration simplifies to only involving inverses of and .

Solutions of algebraic Riccati equations

There is a similar result applicable to the algebraic Riccati equation, .[1][2] Define as

Under the assumption that are Hermitian and there exists a unique stabilizing solution, in the sense that is stable, that solution is given by the over-determined, but consistent, linear system

Proof sketch: The similarity transform

and the stability of implies that

for some matrix .

Computations of matrix square-root

The Denman–Beavers iteration for the square root of a matrix can be derived from the Newton iteration for the matrix sign function by noticing that is a degenerate algebraic Riccati equation[3] and by definition a solution is the square root of .

References

  1. ^ a b c d e f g h Higham, Nicholas J. (2008). Functions of matrices : theory and computation. Society for Industrial and Applied Mathematics. Philadelphia, Pa.: Society for Industrial and Applied Mathematics (SIAM, 3600 Market Street, Floor 6, Philadelphia, PA 19104). ISBN 978-0-89871-777-8. OCLC 693957820.
  2. ^ a b c d Roberts, J. D. "Linear model reduction and solution of the algebraic Riccati equation by use of the sign function". International Journal of Control. 32 (4): 677–687. doi:10.1080/00207178008922881. ISSN 0020-7179.
  3. ^ a b c Denman, Eugene D.; Beavers, Alex N. (1976). "The matrix sign function and computations in systems". Applied Mathematics and Computation. 2 (1): 63–94. doi:10.1016/0096-3003(76)90020-5. ISSN 0096-3003.
  4. ^ Schulz, Günther (1933). "Iterative Berechung der reziproken Matrix". ZAMM - Journal of Applied Mathematics and Mechanics / Zeitschrift für Angewandte Mathematik und Mechanik. 13 (1): 57–59. doi:10.1002/zamm.19330130111. ISSN 1521-4001.