# Lindemann–Weierstrass theorem

In transcendental number theory, the Lindemann–Weierstrass theorem is a result that is very useful in establishing the transcendence of numbers. It states that if α1, ..., αn are algebraic numbers which are linearly independent over the rational numbers ℚ, then eα1, ..., eαn are algebraically independent over ℚ; in other words the extension field ℚ(eα1, ..., eαn) has transcendence degree n over ℚ.

An equivalent formulation (Baker 1975, Chapter 1, Theorem 1.4), is the following: If α1, ..., αn are distinct algebraic numbers, then the exponentials eα1, ..., eαn are linearly independent over the algebraic numbers. This equivalence transforms a linear relation over the algebraic numbers into an algebraic relation over ℚ; by using the fact that a symmetric polynomial whose arguments are all conjugates of one another gives a rational number.

The theorem is named for Ferdinand von Lindemann and Karl Weierstrass. Lindemann proved in 1882 that eα is transcendental for every non-zero algebraic number α, thereby establishing that π is transcendental (see below). Weierstrass proved the above more general statement in 1885.

The theorem, along with the Gelfond–Schneider theorem, is extended by Baker's theorem, and all of these are further generalized by Schanuel's conjecture.

## Naming convention

The theorem is also known variously as the Hermite–Lindemann theorem and the Hermite–Lindemann–Weierstrass theorem. Charles Hermite first proved the simpler theorem where the αi exponents are required to be rational integers and linear independence is only assured over the rational integers,[1] a result sometimes referred to as Hermite's theorem.[2] Although apparently a rather special case of the above theorem, the general result can be reduced to this simpler case. Lindemann was the first to allow algebraic numbers into Hermite's work in 1882.[3] Shortly afterwards Weierstrass obtained the full result,[4] and further simplifications have been made by several mathematicians, most notably by David Hilbert.

## Transcendence of e and π

The transcendence of e and π are direct corollaries of this theorem.

Suppose α is a nonzero algebraic number; then {α} is a linearly independent set over the rationals, and therefore by the first formulation of the theorem {eα} is an algebraically independent set; or in other words eα is transcendental. In particular, e1 = e is transcendental. (A more elementary proof that e is transcendental is outlined in the article on transcendental numbers.)

Alternatively, by the second formulation of the theorem, if α is a nonzero algebraic number, then {0, α} is a set of distinct algebraic numbers, and so the set {e0eα} = {1, eα} is linearly independent over the algebraic numbers and in particular eα cannot be algebraic and so it is transcendental.

The proof that π is transcendental is by contradiction. If π were algebraic, πi would be algebraic as well, and then by the Lindemann–Weierstrass theorem eπi = −1 (see Euler's identity) would be transcendental, a contradiction.

A slight variant on the same proof will show that if α is a nonzero algebraic number then sin(α), cos(α), tan(α) and their hyperbolic counterparts are also transcendental.

p-adic Lindemann–Weierstrass Conjecture. Suppose p is some prime number and α1, ..., αn are p-adic numbers which are algebraic and linearly independent over ℚ, such that | αi |p < 1/p for all i; then the p-adic exponentials expp1), ..., exppn) are p-adic numbers that are algebraically independent over ℚ.

## Modular conjecture

An analogue of the theorem involving the modular function j was conjectured by Daniel Bertrand in 1997, and remains an open problem.[5] Writing q = eiτ for the nome and j(τ) = J(q), the conjecture is as follows. Let q1, ..., qn be non-zero algebraic numbers in the complex unit disc such that the 3n numbers

$\left \{ J(q_1), J'(q_1), J''(q_1), \ldots, J(q_n), J'(q_n), J''(q_n) \right \}$

are algebraically dependent over ℚ. Then there exist two indices 1 ≤ i < j ≤ n such that qi and qj are multiplicatively dependent.

## Lindemann–Weierstrass theorem

Lindemann–Weierstrass Theorem (Baker's Reformulation). If a1, ..., an are non-zero algebraic numbers, and α1, ..., αn are distinct algebraic numbers, then[6]

$a_1 e^{\alpha_1} +\cdots + a_n e^{\alpha_n}\ne 0.$

### Proof

#### Preliminary lemmas

Lemma A. Let c(1), ..., c(r) be non-zero integers and, for every k between 1 and r, let {γ(k)1, ..., γ(k)m(k)} be the roots of a polynomial with integer coefficients Tk (x) = v(k) xm(k) + ... + u(k) (with v(k), u(k) ≠ 0). If γ(k)i ≠ γ(u)v whenever (ki) ≠ (uv), then

$c(1)\left (e^{\gamma(1)_1}+\cdots+ e^{\gamma(1)_{m(1)}} \right ) + \cdots + c(r) \left (e^{\gamma(r)_1}+\cdots+ e^{\gamma(r)_{m(r)}} \right) \ne 0.$

Proof of Lemma A. To simplify the notation set:

$n_0=0, \qquad n=n_r$
$n_i=\sum\nolimits_{k=1}^i m(k) \qquad i=1,\ldots,r$
$\alpha_{n_i+j}=\gamma(i+1)_j \qquad 0\leqslant i
$\beta_{n_i+j}=c(i+1)$

Then the statement becomes:

$\sum_{k=1}^n \beta_k e^{\alpha_k}\neq 0.$

Let p be a prime number and define the following polynomials:

$f_i(x) = \frac {l^{np} (x-\alpha_1)^p \cdots (x-\alpha_n)^p}{(x-\alpha_i)},$

where l is a non-zero integer such that $l\alpha_1,\ldots,l\alpha_n$ are all algebraic integers. Define:[7]

$I_i(s) = \int^s_0 e^{s-x} f_i(x) \, dx.$

Using integration by parts we arrive at:

$I_i(s) = e^s \sum_{j=0}^{np-1} f_i^{(j)}(0) - \sum_{j=0}^{np-1} f_i^{(j)}(s),$

where $np-1$ is the degree of $f_i$, and $f_i^{(j)}$ is the j-th derivative of $f_i$. This also holds for s complex (in this case the integral has to be intended as a contour integral, for example along the straight segment from 0 to s) because

$-e^{s-x} \sum_{j=0}^{np-1} f_i^{(j)}(x)$

is a primitive of $e^{s-x} f_i(x)$.

Let us consider the following sum:

\begin{align} J_i &=\sum_{k=1}^n\beta_k I_i(\alpha_k)\\ &= \sum_{k=1}^n\beta_k \left ( e^{\alpha_k} \sum_{j=0}^{np-1} f_i^{(j)}(0) - \sum_{j=0}^{np-1} f_i^{(j)}(\alpha_k)\right ) \\ &=\sum_{k=1}^n\left(\beta_k e^{\alpha_k}\sum_{j=0}^{np-1}f_i^{(j)}(0)\right)-\sum_{k=1}^n\left(\beta_k\sum_{j=0}^{np-1}f_i^{(j)}(\alpha_k)\right) \\ &=\left(\sum_{j=0}^{np-1}f_i^{(j)}(0)\right)\left(\sum_{k=1}^n \beta_k e^{\alpha_k}\right)-\sum_{k=1}^n\sum_{j=0}^{np-1} \beta_kf_i^{(j)}(\alpha_k)\\ &= -\sum_{k=1}^n \sum_{j=0}^{np-1} \beta_kf_i^{(j)}(\alpha_k) && \sum_{k=1}^n \beta_k e^{\alpha_k}=0 \end{align}

In the last line we assumed that the conclusion of the Lemma is false. In order to complete the proof we need to reach a contradiction. We will do so by estimating $|J_1\cdots J_n|$ in two different ways.

First $f_i^{(j)}(\alpha_k)$ is an algebraic integer which is divisible by p! for $j\geqslant p$ and vanishes for j < p unless j=p-1 and k=i, in which case it equals

$l^{np}(p-1)!\prod_{k\neq i}(\alpha_i-\alpha_k)^p$.

This is not divisible by p (if p is large enough) because otherwise, putting

$\delta_i=\prod_{k\neq i}(l\alpha_i-l\alpha_k)$

(which is an algebraic integer) and calling $d_i$ the product of its conjugates, we would get that p divides $l^p(p-1)!d_i^p$ (and $d_i$ is a non-zero integer), so by Fermat's little theorem p would divide $l(p-1)!d_i$, which is false.

So $J_i$ is a non-zero algebraic integer divisible by (p-1)!. Now

$J_i=-\sum_{j=0}^{np-1}\sum_{t=0}^{r-1}c(t+1)\left(f_i^{(j)}(\alpha_{n_t+1})+\cdots+f_i^{(j)}(\alpha_{n_{t+1}})\right).$

Since each $f_i(x)$ is obtained by dividing a fixed polynomial with integer coefficients by $(x-\alpha_i)$, it is of the form

$f_i(x)=\sum_{m=0}^{np-1}g_m(\alpha_i)x^m,$

where $g_m$ is a polynomial (with integer coefficients) independent of i. The same holds for the derivatives $f_i^{(j)}(x)$.

Hence, by the fundamental theorem of symmetric polynomials,

$f_i^{(j)}(\alpha_{n_t+1})+\cdots+f_i^{(j)}(\alpha_{n_{t+1}})$

is a fixed polynomial with integer coefficients evaluated in $\alpha_i$ (this is seen by grouping the same powers of $\alpha_{n_t+1},\dots,\alpha_{n_{t+1}}$ appearing in the expansion and using the fact that $\alpha_{n_t+1},\cdots,\alpha_{n_{t+1}}$ are a complete set of conjugates). So the same is true of $J_i$, i.e. it equals $G(\alpha_i)$, where G is a polynomial with integer coefficients which is independent of i.

Finally $J_1\dots J_n=G(\alpha_1)\dots G(\alpha_n)$ is an integer number (again by the fundamental theorem of symmetric polynomials), it is non-zero (since the $J_i$'s are) and it is divisible by $(p-1)!^n$ and therefore:

$|J_1\cdots J_n|\geqslant (p-1)!^n.$

However one clearly has:

$|I_i(\alpha_k)| \leqslant |\alpha_k|e^{|\alpha_k|}F_i(|\alpha_k|),$

where Fi is the polynomial whose coefficients are the absolute values of those of fi (this follows directly from the definition of $I_i(s)$). Thus

$|J_i|\leqslant \sum_{k=1}^n \left |\beta_k\alpha_k \right |e^{|\alpha_k|}F_i \left ( \left |\alpha_k \right| \right )$

and so by the construction of the $f_i$'s we have $|J_1\cdots J_n|\le C^p$ for a sufficiently large C independent of p, which contradicts the previous inequality. This proves Lemma A.

Lemma B. If b(1), ..., b(n) are non-zero integers and γ(1), ..., γ(n), are distinct algebraic numbers, then

$b(1)e^{\gamma(1)}+\cdots+ b(n)e^{\gamma(n)}\ne 0.$

Proof of Lemma B: Assuming

$b(1)e^{\gamma(1)}+\cdots+ b(n)e^{\gamma(n)}= 0,$

we will derive a contradiction, thus proving Lemma B.

Let us choose a polynomial with integer coefficients which vanishes on all the $\gamma(k)$'s and let $\gamma(1),\ldots,\gamma(n),\gamma(n+1),\ldots,\gamma(N)$ be all its distinct roots. Let b(n + 1) = ... = b(N) = 0.

The product

$\prod_{\sigma\in S_N}(b(1) e^{\gamma(\sigma(1))}+\cdots+b(N) e^{\gamma(\sigma(N))}),$

vanishes by assumption, but by expanding it we obtain a sum of terms of the form $e^{h_1\gamma(1)+\cdots+h_N\gamma(N)}$ multiplied by integer coefficients. Since the product is symmetric, we have that, for any $\tau\in S_n$, $e^{h_1\gamma(\tau(1))+\cdots+h_N\gamma(\tau(N))}$ has the same coefficient as $e^{h_1\gamma(1)+\dots+h_N\gamma(N)}$.

Thus (after having grouped the terms with the same exponent) we see that the set of the exponents form a complete set of conjugates and, if two terms have conjugate exponents, they are multiplied by the same coefficient.

So we are in the situation of Lemma A. To reach a contradiction it suffices to see that at least one of the coefficients is non-zero. This is seen by equipping C with the lexicographic order and by choosing for each factor in the product the term with non-zero coefficient which has maximum exponent according to this ordering: the product of these terms has non-zero coefficient in the expansion and does not get simplified by any other term. This proves Lemma B.

#### Final step

We turn now to prove the theorem: Let a(1), ..., a(n) be non-zero algebraic numbers, and α(1), ..., α(n) distinct algebraic numbers. Then let us assume that:

$a(1)e^{\alpha(1)}+\cdots + a(n)e^{\alpha(n)} = 0.$

We will show that this leads to contradiction and thus prove the theorem.

The proof is very similar to that of Lemma B, except that this time the choices are made over the a(i)'s:

For every i ∈ {1, ..., n}, a(i) is algebraic, so it is a root of a polynomial with integer coefficients, we denote its degree by d(i). Let us denote the roots of this polynomial a(i)1, ..., a(i)d(i), with a(i)1 = a(i).

Let σ be a function which chooses one element from each of the sequences (1, ..., d(1)), (1, ..., d(2)), ..., (1, ..., d(n)), such that for every 1 ≤ i ≤ n, σ(i) is an integer between 1 and d(i). Then according to our assumption:

$\prod\nolimits_{\{\sigma\}}\left(a(1)_{\sigma(1)}e^{\alpha(1)}+\cdots+ a(n)_{\sigma(n)} e^{\alpha(n)}\right) = 0$

where the product is over all possible choices. The product vanishes because one of the choices is just σ(i) = 1 for all i, for which the term vanishes according to our assumption above.

By expanding this product we get a sum of the form:

$b(1)e^{\beta(1)}+ b(2)e^{\beta(2)}+ \cdots + b(N)e^{\beta(N)}= 0.$

for some non-zero integer N, some distinct algebraic β(1), ..., β(N) (these are indeed algebraic because each is a sum of α's which are algebraic themselves), and b(1), ..., b(N) are polynomial in a(i)j (i in 1, ..., n and j in 1, ..., d(i)) with integer coefficients.

Since the product is over all possible choices, each of b(1), ..., b(N) is symmetric in a(i)1, ..., a(i)d(i) for every i; therefore each of b(1), ..., b(N) is a polynomial with integer coefficients in elementary symmetric polynomials of the sets {a(i)1, ..., a(i)d(i)} for every i. Each of the latter is a rational number (as in the proof of Lemma B).

Thus b(1), ..., b(N) ∈ Q, and by multiplying the equation with an appropriate integer factor, we get an identical equation except that now b(1), ..., b(N) are all integers.

Therefore, according to Lemma B, the equality cannot hold, and we are led to a contradiction which completes the proof.

Note that Lemma A is sufficient to prove that π is irrational, since otherwise we may write π = k/n (kn, integers) and then ±iπ are the roots of x2 + k2/n2; thus 2 + eiπ + eiπ ≠ 0; but this is false.

Similarly, Lemma B is sufficient to prove that π is transcendental, since otherwise we would have 1 + eiπ ≠ 0.

## References

1. ^ Sur la fonction exponentielle, Comptes Rendus Acad. Sci. Paris, 77, pages 18–24, 1873.
2. ^ A.O.Gelfond, Transcendental and Algebraic Numbers, translated by Leo F. Boron, Dover Publications, 1960.
3. ^ Über die Ludolph'sche Zahl, Sitzungsber. Königl. Preuss. Akad. Wissensch. zu Berlin, 2, pages 679–682, 1882.
4. ^ Zu Hrn. Lindemanns Abhandlung: 'Über die Ludolph'sche Zahl' , Sitzungber. Königl. Preuss. Akad. Wissensch. zu Berlin, 2, pages 1067–1086, 1885
5. ^ Daniel Bertrand, Theta functions and transcendence, The Ramanujan Journal 1, pages 339–350, 1997.
6. ^
7. ^ Up to a factor, this is the same integral appearing in the proof that e is a transcendental number, where β1 = 1, ..., βm= m. The rest of the proof of the Lemma is analog to that proof.