Mathematically vectors are elements of a vector space over a field , and for use in physics is usually defined with or . Concretely, if the dimension of is finite, then, after making a choice of basis, we can view such vector spaces as or .
The dual space is the space of linear functionals mapping . Concretely, in matrix notation these can be thought of as row vectors, which give a number when applied to column vectors. We denote this by , so that is a linear map .
Then under a choice of basis , we can view vectors as an vector with components (vectors are taken by convention to have indices up). This picks out a choice of basis for , defined by the set of relations .
For applications, raising and lowering is done using a structure known as the (pseudo-)metric tensor (the 'pseudo-' refers to the fact we allow the metric to be indefinite). Formally, this is a non-degenerate, symmetric bilinear form
In this basis, it has components , and can be viewed as a symmetric matrix in with these components. The inverse metric exists due to non-degeneracy and is denoted , and as a matrix is the inverse to .
Raising and lowering vectors and covectors
Raising and lowering is then done in coordinates. Given a vector with components , we can contract with the metric to obtain a covector:
and this is what we mean by lowering the index. Conversely, contracting a covector with the inverse metric gives a vector:
This process is called raising the index.
Raising and then lowering the same index (or conversely) are inverse operations, which is reflected in the metric and inverse metric tensors being inverse to each other (as is suggested by the terminology):
Finite-dimensional real vector spaces with (pseudo-)metrics are classified up to signature, a coordinate-free property which is well-defined by Sylvester's law of inertia. Possible metrics on real space are indexed by signature . This is a metric associated to dimensional real space. The metric has signature if there exists a basis (referred to as an orthonormal basis) such that in this basis, the metric takes the form with positive ones and negative ones.
The concrete space with elements which are -vectors and this concrete realization of the metric is denoted , where the 2-tuple is meant to make it clear that the underlying vector space of is : equipping this vector space with the metric is what turns the space into .
Examples:
is a model for 3-dimensional space. The metric is equivalent to the standard dot product.
, equivalent to dimensional real space as an inner product space with . In Euclidean space, raising and lowering is not necessary due to vectors and covector components being the same.
is Minkowski space (or rather, Minkowski space in a choice of orthonormal basis), a model for spacetime with weak curvature. It is common convention to use greek indices when writing expressions involving tensors in Minkowski space, while Latin indices are reserved for Euclidean space.
Well-formulated expressions are constrained by the rules of Einstein summation: any index may appear at most once and furthermore a raised index must contract with a lowered index. With these rules we can immediately see that an expression such as
To raise the index, multiply by the tensor and contract:
then for λ = 0:
and for λ = j = 1, 2, 3:
So the index-raised contravariant 4-position is:
This operation is equivalent to the matrix multiplication
Given two vectors, and , we can write down their (pseudo-)inner product in two ways:
By lowering indices, we can write this expression as
What is this in matrix notation? The first expression can be written as
while the second is, after lowering the indices of ,
Coordinate free formalism
It is instructive to consider what raising and lowering means in the abstract linear algebra setting.
We first fix definitions: is a finite-dimensional vector space over a field . Typically or .
is a non-degenerate bilinear form, that is,
is a map which is linear in both arguments, making it a bilinear form.
By being non-degenerate we mean that for each , there is a such that
In concrete applications, is often considered a structure on the vector space, for example an inner product or more generally a metric tensor which is allowed to have indefinite signature, or a symplectic form . Together these cover the cases where is either symmetric or anti-symmetric, but in full generality need not be either of these cases.
There is a partial evaluation map associated to ,
where denotes an argument which is to be evaluated, and denotes an argument whose evaluation is deferred. Then is an element of , which sends .
We made a choice to define this partial evaluation map as being evaluated on the first argument. We could just as well have defined it on the second argument, and non-degeneracy is also independent of argument chosen. Also, when has well defined (anti-)symmetry, evaluating on either argument is equivalent (up to a minus sign for anti-symmetry).
Non-degeneracy shows that the partial evaluation map is injective, or equivalently that the kernel of the map is trivial. In finite dimension, the dual space has equal dimension to , so non-degeneracy is enough to conclude the map is a linear isomorphism. If is a structure on the vector space sometimes call this the canonical isomorphism .
It therefore has an inverse, and this is enough to define an associated bilinear form on the dual:
where the repeated use of is disambiguated by the argument taken. That is, is the inverse map, while is the bilinear form.
Checking these expressions in coordinates makes it evident that this is what raising and lowering indices means abstractly.
Tensors
We will not develop the abstract formalism for tensors straightaway. Formally, an tensor is an object described via its components, and has components up, components down. A generic tensor is written
We can use the metric tensor to raise and lower tensor indices just as we raised and lowered vector indices and raised covector indices.
Examples
A (0,0) tensor is a number in the field .
A (1,0) tensor is a vector.
A (0,1) tensor is a covector.
A (0,2) tensor is a bilinear form. An example is the metric tensor
A (1,1) tensor is a linear map. An example is the delta, , which is the identity map, or a Lorentz transformation
Example of raising and lowering
For a (0,2) tensor,[1] twice contracting with the inverse metric tensor and contracting in different indices raises each index:
Similarly, twice contracting with the metric tensor and contracting in different indices lowers each index:
Let's apply this to the theory of electromagnetism.
To obtain the covariant tensor Fαβ, contract with the inverse metric tensor:
and since F00 = 0 and F0i = − Fi0, this reduces to
Now for α = 0, β = k = 1, 2, 3:
and by antisymmetry, for α = k = 1, 2, 3, β = 0:
then finally for α = k = 1, 2, 3, β = l = 1, 2, 3;
The (covariant) lower indexed tensor is then:
This operation is equivalent to the matrix multiplication
General rank
For a tensor of order n, indices are raised by (compatible with above):[1]
and lowered by:
and for a mixed tensor:
We need not raise or lower all indices at once: it is perfectly fine to raise or lower a single index. Lowering an index of an tensor gives a tensor, while raising an index gives a (where have suitable values, for example we cannot lower the index of a tensor.)
^ abKay, D. C. (1988). Tensor Calculus. Schaum’s Outlines. New York: McGraw Hill. ISBN0-07-033484-6.
^NB: Some texts, such as: Griffiths, David J. (1987). Introduction to Elementary Particles. Wiley, John & Sons, Inc. ISBN0-471-60386-4., will show this tensor with an overall factor of −1. This is because they used the negative of the metric tensor used here: (− + + +), see metric signature. In older texts such as Jackson (2nd edition), there are no factors of c since they are using Gaussian units. Here SI units are used.