Talk:Dimension theorem for vector spaces

From Wikipedia, the free encyclopedia
Jump to: navigation, search
WikiProject Mathematics (Rated Start-class, Low-importance)
WikiProject Mathematics
This article is within the scope of WikiProject Mathematics, a collaborative effort to improve the coverage of Mathematics on Wikipedia. If you would like to participate, please visit the project page, where you can join the discussion and see a list of open tasks.
Mathematics rating:
Start Class
Low Importance
 Field: Algebra

Could someone write about how to prove the uniqueness of the dimension for infinite cardinals? - Gauge 06:02, 18 December 2005 (UTC)

I know its been a year and a half since anyone replied, but done. jbolden1517Talk 15:19, 12 May 2007 (UTC)


I think the whole thing with going back and forth between infinite and finite violates the spirit of cardinality. It should be a single case. Actually looking at it the infinite case is false. jbolden1517Talk 17:48, 12 May 2007 (UTC)

I do not know if there is a nice proof that works for both the finite and infinite case. I do not agree that a case distinction violates the "spirit of cardinality". Finite cardinals behave very differently from infinite one, this is why there are two (often quite separate) fields of mathematics dealing with them, combinatorics and set theory.
You deleted my argument for the case of finite dimensional spaces. But the argument
"Since the set J has smaller cardinality than the set I, the union of finite sets indexed by J still has smaller cardinality than I"
works only if I is infinite.
--Aleph4 00:18, 27 May 2007 (UTC)

I changed your proof a lot. The first line just asserts that as the RAA hypothesis. So its true by given. jbolden1517Talk 00:38, 27 May 2007 (UTC)

What do you mean by "RAA"? Which first line asserts what?
Please explain why the cardinality of I is bigger than the cardinality of the union of the E_j, j in J.
--Aleph4 14:45, 29 May 2007 (UTC)

I don't really know where to write this, but this proof is wrong, and quite obviously so, too :
just consider the case where b_1 is the sum of the a_i. Then E_1 is I, and there's no i_0 such that ...

This theorem is actually not so easy to prove.
I'm not sure there exist a simple elegant proof, that's what I was looking for here.
The proof I know is no fun.
I also know it can be seen as a consequence of the Jordan-Holder theorem (see Bourbaki), but I ghess that would be overkill —Preceding unsigned comment added by (talk) 15:45, 17 February 2008 (UTC)

I again divided the proof into two cases, one for finite dimension, the other for infinite dimension. As I have pointed out before, the proof from the article that works for infinite dimensions does not work in the finite-dimensional case. --

Aleph4 (talk) 17:44, 17 February 2008 (UTC)

The proof for the infinite case does indeed look wrong (, as someone mentioned). Could someone replace it with a correct proof? If not, it should be removed.

Don't know when the above was written, but it seems correct now, the sum is over a finite set for every b. However, the supposition that one of the bases has greater cardinal tan the other depends on the (full) axiom of choice, contrary to what the lead suggests. Marc van Leeuwen (talk) 11:57, 20 March 2011 (UTC)

AC in Proof?[edit]

In the proof it is assumed that there is a relation between the cardinalities of the two bases (that one is greater than the other). It seems to me that this relies on the Well-Ordering Principle. However I'm not familiar with the Ultra-filter Lemma and its applications... is this where it is used? There is a proof here: that relies only on propositional compactness and Hall's Matching Theorem.

You (who?) are right to the extent that trichotomy (any two cardinals are comparable) is equivalent to the axiom of choice, as stated in that article. So indeed the very start of the proof (assume one basis bigger than the other) is in blatant contradiction with the claim in the lead that the theorem only needs the ultrafilter lemma. Marc van Leeuwen (talk) 05:43, 20 March 2011 (UTC)
Although this particular proof does use full AC, it is nevertheless true that the dimension theorem (also known as Löwig's theorem) follows from the ultrafilter lemma. I have checked the primary source on this kind of issues: "Consequences of the axiom of choice" by Howard & Rubin. You can also check it online at .Godelian (talk) 05:09, 11 October 2011 (UTC)
I've put up a somewhat rough proof at ProofWiki. That proof uses Hall's marriage theorem for infinite sets of finite sets, which I prove here from the finite form of the marriage theorem along with the Cowen-Engeler Lemma, an equivalent of the Ultrafilter Lemma. --Dfeuer (talk) 04:04, 24 May 2013 (UTC)
Please put a (more detailed) reference in the article as well. Thanks a lot. Marc van Leeuwen (talk) 08:45, 11 October 2011 (UTC)
Done. Godelian (talk) 23:57, 11 October 2011 (UTC)

Possible error?[edit]

I'm probably wrong, but in the alternative proof (the only one I understand): "there is at least one such that can be written as a linear combination of . Thus, ."

I agree with the last sentence, however to be technically correct, I believe that "there is at least one such that can be written as a linear combination of . Thus..."

Am I wrong? If so, please delete this. Spartan S58 (talk) 08:18, 1 October 2011 (UTC)

Your alternative formulation is easier to understand, but the one currently in place is (almost) correct, because if one takes the smallest t such that the family is linearly dependent (and it exists because for one has dependence) then by minimality the final vector is a linear combination of the preceding ones. However this somewhat stronger statement does not appear to be used in the proof, so I agree that morally the somewhat longer statement should have been there. Yet the current proof is clumsy, and IMHO should be rewritten anyway. What should be used is what in France we call the fr:Théorème de la base incomplète (don't know if it has an English name), which I like to state in the following form: if F is a free (possibly empty) set of vectors and S a spanning list, then there exist a basis consisting of F together with a subset of S; moreover with the additional requirement the every retained element s of S be linearly independent on F united with the elements of S (strictly) before s, this basis is unique. The proof is by induction on the size of S: if one has selected the unique basis of (with the additional requirement), then either already , and then the additional condition forbids including in the basis so one must take as basis for V, or else W is strictly smaller than V, and since together with it spans V, W cannot contain ; then our basis for V must include , and the remainder is forced to be , and indeed is a basis of V satisfying the additional condition.
And to justify my "(almost) correct" above: when the current proof talks about , it forgets that removing could also inadvertently remove (when ). It should have said . Marc van Leeuwen (talk) 07:24, 10 October 2011 (UTC)
A small correction to what I said above. I realised that as stated the inductive argument is wrong because the assumption is not satisfied when the induction hypothesis is applied: need not generate . The solution is surprisingly simple: for the sake of the inductive proof, weaken the condition in the statement from "S a spanning list" to "S is a list such that generates V"; the more general statement remains true, and in fact the inductive proof becomes valid (this is a case of Strengthening the induction hypothesis, for which I would have expected to find a WP article). Also, since I brought up the point that F and S could overlap, I should specify that a "retained element" of S is one that is in the basis but is not in F (without the latter clause, the elements of would necessarily be retained, but they cannot satisfy the additional requirement). Marc van Leeuwen (talk) 09:27, 10 October 2011 (UTC)
I must admit your proof loses me. I am a freshman in college, just turned 18, with an Intro to Lin Alg class Spring 2011, my last semester of high school. On the other hand, having reviewed the link, I remember that the proof of the dimension theorem that I saw in the class I took involved either cutting down a spanning but dependent set to a basis, or constructing a basis from an independent set that didn't span; this seems to be about what the French theorem talks about, just that it's a separate theorem from the dimension theorem. Also, having read that, I think part of my confusion with your proof initially comes from your use of the word "free" to mean linearly independent. I've never seen free used in that sense in English, though clearly in the French article. I give your proof another shot. (Also I have no clue what you mean by minimality, regarding my original question.) ((In an unrelated note, mijn Opa (en Oma) are from the Netherlands, but I don't speak Dutch :P)) Spartan S58 (talk) 22:49, 17 October 2011 (UTC)