I removed the condition that the matrix be square for the induced norm (when p = 2) to be equivalent to the largest singular value. Indeed, this equivalence is true for non-square matrices too.
The following page will be replaced by a table.--wshun 01:34, 8 Aug 2003 (UTC)
The most "natural" of these operator norms is the one which arises from the Euclidean norms ||.||2 on Km and Kn. It is unfortunately relatively difficult to compute; we have
(see singular value). If we use the taxicab norm ||.||1 on Km and Kn, then we obtain the operator norm
and if we use the maximum norm ||.||∞ on Km and Kn, we get
The following inequalities obtain among the various discussed matrix norms for the m-by-n matrix A:
- 1 German Link
- 2 What's wrong with Frobenius norm?
- 3 What happened to the article?
- 4 Induced norm
- 5 Submultiplicativity
- 6 Bad Notation
- 7 Matrix Norm not Vector Norm
- 8 equivalence of norms
- 9 trace norm vs. Frobenius norm
- 10 trace norm
- 11 Gradient of the Norm
- 12 Matrix Norm Definition
- 13 Matrix Norm Example
- 14 max?
- 15 spectral radius
- 16 Thank you
- 17 Centralized discussion on proofs
- 18 Poor article layout
This site needs to be linked to http://de.wikipedia.org/wiki/Matrixnorm
What's wrong with Frobenius norm?
Why does the article say that Frobenius norm is not sub-multiplicative? It does satisfy the condition , which can be easily proved as follows: . --Igor 21:21, Feb 18, 2005 (UTC)
Is it true that the Frobenius norm is when p=2. It seems to me that it is the norm that is mentioned earlier in the article. . Also it is also called the Hilbert-Schmidt norm, because the page for Hilbert-Schmidt norm says that it is only analogous to the Frobenius norm.--kfrance 13:40, Oct 9, 2007 (MST)
- @KFrance, That is not true. The Frobenius norm is the Hilbert-Schmidt norm, but it is not the same as (this is the 'spectral norm'). For vectors, is the Euclidean norm which is the same as the Frobenius norm if the input vector is treated like a matrix, but when the input is a matrix, the notation usually denotes spectral norm, which is not the Frobenius norm. @Igor, that is true. Lavaka (talk) 17:54, 9 July 2014 (UTC)
What happened to the article?
The above discussion suggests that the article used to be more extensive. However, the revision history of the current article shows only one edit, by CyborgTosser on 25 Feb 2005. Did something drastic happen to the article? -- Jitse Niesen 11:36, 2 Mar 2005 (UTC)
- I'm not quite sure what happened. Apparently there used to be an article here, but the content must have been moved. I'm not sure where and I'm not sure why, but a lot of articles link here, so I figured we needed the article. Hopefully whoever moved the content will replace whatever is relevant. CyborgTosser (Only half the battle) 03:21, 11 Mar 2005 (UTC)
- I don't know either. I couldn't find the old page on wikipedia with google, but I've put a copy (from a wikipedia clone) at Matrix norm/old. Lupin 13:50, 11 Mar 2005 (UTC)
- It seems that User:RickK deleted this page after it had been vandalised. Idiot. I've asked him to restore it with edit history to a subpage if possible. Lupin 14:10, 11 Mar 2005 (UTC)
I'm a little confused where the article says that "any induced norm satisfies the inequality ...". Is the intended meaning that the operator norm satisfies that inequality, or are there other norms which are also known as induced norms which satisfy that inequality? If the former, it should be rephrased as "the induced norm satisfies..." and if the latter, an explanation of what is meant by an induced norm should be given. Lupin 01:24, 11 May 2005 (UTC)
- The terms "induced norm" and "operator norm" are synonymous. I used "any induced norm" instead of "the induced norm" because there are several operator norms. One example is the spectral norm, another example arises when one takes the ∞-norm on Kn, defined by
- the resulting operator norm is
- I hope this resolves the confusion; feel free (of course) to edit the article to make it clearer. -- Jitse Niesen 10:23, 11 May 2005 (UTC)
I feel that this article is quite unclear about when submultiplicativity applies. In particular, it should be made clear that for matrix norms based on vectors p-norms that for and that . This is shown in Proposition 2.7.2 on the following page .
- You are right that this could be added. So, why don't you change the article to include this? You can edit the article by clicking on "edit this page", see How to edit a page for details. Don't worry about making mistakes; you will be corrected if necessary. I look forward to your contributions, Jitse Niesen (talk) 11:24, 12 August 2005 (UTC)
Moreover, when m = n, then for any vector norm | · |, there exists a unique positive number k such that k| · | is a (submultiplicative) matrix norm. A matrix norm || · || is said to be minimal if there exists no other matrix norm | · | satisfying |A|≤||A|| for all |A|.
Doesn't |A| specify the absolute value? Using the correct notation yields ||A||≤||A|| for all ||A||. Isn't that self evident? Furthermore m and n are not specified. Therefore I have removed this section till someone can clarify this content. It looks as if though someone partially moved content such that it's meaning was lost. —The preceding unsigned comment was added by ANONYMOUS COWARD0xC0DE (talk • contribs) 02:53, 24 December 2006 (UTC).
So sorry; don't know what I was thinking. I will just change |A| to ||A||_q and ||A|| to ||A||_p, it's clear from the sentence what |A| refereed to. I was reading a book earlier and |A| was refereed to as the determinant of A. More-over I will just add these statements back in and reword them. --ANONYMOUS COWARD0xC0DE 01:06, 29 December 2006 (UTC)
Matrix Norm not Vector Norm
* * * * * *
These are properties of vectors of the form and not of the form . --ANONYMOUS COWARD0xC0DE 03:38, 24 December 2006 (UTC)
equivalence of norms
article is not really clear about the equivalence of norms: since we are talking about matrices of finite size, all vector norms should be equivalent. the bunch of inequalities in the bottom could (mis)lead the reader into thinking otherwise. if, in addition, submultiplicativity is required, does this change? (apparently so, the article seems to imply the Banach algebra topology is not unique.) Mct mht 14:08, 13 February 2007 (UTC)
trace norm vs. Frobenius norm
it isn't true that the trace norm, sum(sigma), is <= the Frob. norm, sum(sigma^2); e.g. suppose all sigma<1. Lpwithers 16:34, 8 October 2007 (UTC)
Gradient of the Norm
I'm interested in learning about the gradient of the matrix norm but I can't seem to find this information within wikipedia. I guess I'm requesting a new article and I don't know where to do that, but it seems logical for this article to point me to the gradient of the norm (maybe under see also). —Preceding unsigned comment added by Arbitrary18 (talk • contribs) 01:00, 23 September 2008 (UTC)
Matrix Norm Definition
Matrix norm on the set of all nxn matrices is a real value function, ||.|| defined on this set, satisfying for all nxn matrices A and B and all real number :
- if and if and only if
- for all in and all matrices in
- for all matrices and in
Matrix Norm Example
The two following functions are two examples of matrix norm:
For examples: With matrix A:
Note: In the above example is the maximum absolute column sum of the matrix, and is the maximum absolute row sum of the matrix. In addition both and are the special norm of a general norm called p-norm for vectors
In some of the definitions I wasn't sure if max should actually be the supremum. I thought a maximum is guaranteed to exist for compact sets of real numbers, but not necessarily for open sets. In the case of linear, finite-dimensional operators(open sets are mapped to open sets) wouldn't this be equivalent to the domain being compact? In the case of the induced norm that would imply (from my perspective) max in the case abs(x)<=1 and supremum in the case x not equal to zero. I am not sure if it is actually an issue or not because at least in case of the induced 2 norm, the supremum is actually part of the range. That in turn implies to me that the supremum is reached for any similarly defined induced norm because of the equivalence of norms in finite dimensional spaces. Can someone with experience maybe point out the disconnect I seem to be having? —Preceding unsigned comment added by 126.96.36.199 (talk) 18:49, 19 July 2010 (UTC)
- The domain is usually a sphere. These are closed and bounded, and thus compact by Heine-Borel. — Preceding unsigned comment added by 188.8.131.52 (talk) 17:06, 1 August 2011 (UTC)
There is a statement in the article: "For a symmetric or hermitian matrix A, we have equality for the 2-norm, since in this case the 2-norm is the spectral radius of A" I guess the equality actually holds for more general case: It holds for any diagonalizable A. (Note that symmetric/hermitian is a special case of diagonalizable matrices when the diagonalizing matrix are unitary, which in turn, is a special case of normal matrices. All these are diagonalizable.) Trivial proof: Let A = P D P-1. Then (since the set of eigenvalues of AB is same as the set of eigenvalue of BA)
- That argument was wrong. If gives the set of eigenvalues of matrix , then if is a cyclic permutation. - Subh83 (talk | contribs) 04:23, 8 February 2013 (UTC)
Centralized discussion on proofs
Poor article layout
It would be much clearer if the definitions of the norms and their properties was more clearly demarcated. At present, being sub-multiplicative is defined at the top, but the fact that all induced norms are sub-multiplicative is just mentioned in passing in the discussion of induced norms. Contrast this to consistency, for which the fact that induced norms are consistent is mentioned next to the definition of consistency.
I would suggest one of the following two layouts:
- Start with the definition of a matrix norm, and the formal definition of each property that such a norm might have. Then go through the definitions of induced, Frobenius etc. norms, with clear results for each norm on which properties it does (not) possess.
- Start with the definition of a matrix norm, then go through the definitions of induced, Frobenius etc. norms as examples. Then go through the definitions of each property matrix norms might have, with clear results on which norms (do not) possess the given property.
In either approach, a table of norms and properties might help the presentation.