Tensor (intrinsic definition)
In mathematics, the modern component-free approach to the theory of a tensor views a tensor as an abstract object, expressing some definite type of multi-linear concept. Their well-known properties can be derived from their definitions, as linear maps or more generally; and the rules for manipulations of tensors arise as an extension of linear algebra to multilinear algebra.
In differential geometry an intrinsic geometric statement may be described by a tensor field on a manifold, and then doesn't need to make reference to coordinates at all. The same is true in general relativity, of tensor fields describing a physical property. The component-free approach is also used heavily in abstract algebra and homological algebra, where tensors arise naturally.
- Note: This article assumes an understanding of the tensor product of vector spaces without chosen bases. An overview of the subject can be found in the main tensor article.
Definition via tensor products of vector spaces
A tensor on the vector space V is then defined to be an element of (i.e., a vector in) a vector space of the form:
where V∗ is the dual space of V.
If there are m copies of V and n copies of V∗ in our product, the tensor is said to be of type (m, n) and contravariant of order m and covariant order n and total order m + n. The tensors of order zero are just the scalars (elements of the field F), those of contravariant order 1 are the vectors in V, and those of covariant order 1 are the one-forms in V∗ (for this reason the last two spaces are often called the contravariant and covariant vectors). The space of all tensors of type (m, n) is denoted
The type (1, 1) tensors
The term rank of a tensor extends the notion of the rank of a matrix in linear algebra, although the term is also often used to mean the order (or degree) of a tensor. The rank of a matrix is the minimum number of column vectors needed to span the range of the matrix. A matrix thus has rank one if it can be written as an outer product of two nonzero vectors:
More generally, the rank of a matrix A is the smallest number of such outer products that can be summed to produce it:
Similarly, a tensor of rank one (also called a simple tensor) is a tensor that can be written as a tensor product of the form
where a, b, ..., d are nonzero and in V or V*. That is, if the tensor is nonzero and completely factorizable. In indices, a tensor of rank 1 is a tensor of the form
Every tensor can be expressed as a sum of rank 1 tensors. The rank of a general tensor T is defined to be the minimum number of rank 1 tensors with which it is possible to express T as a sum (Bourbaki 1989, II, §7, no. 8).
A nonzero order 1 tensor always has rank 1. The zero tensor has rank zero. The rank of a non-zero order 2 or higher tensor is less than or equal to the product of all but the longest of the lengths of the vectors in (a sum of products of) which the tensor can be expressed. The rank of a tensor of order 2 agrees with the rank when the tensor is regarded as a matrix (Halmos 1974, §51), and can be determined from Gaussian elimination for instance. The rank of an order 3 or higher tensor is however often very hard to determine, and low rank decompositions of tensors are sometimes of great practical interest (de Groote 1987). Computational tasks such as the efficient multiplication of matrices and the efficient evaluation of polynomials can be recast as the problem of simultaneously evaluating a set of bilinear forms
The space can be characterized by a universal property in terms of multilinear mappings. Amongst the advantages of this approach are that it gives a way to show that many linear mappings are "natural" or "geometric" (in other words are independent of any choice of basis). Explicit computational information can then be written down using bases, and this order of priorities can be more convenient than proving a formula gives rise to a natural mapping. Another aspect is that tensor products are not used only for free modules, and the "universal" approach carries over more easily to more general situations.
is multilinear if it is linear in each argument. The space of all multlinear mappings from the product V1×V2×...×VN into W is denoted LN(V1,V2,...,VN; W). When N = 1, a multilinear mapping is just an ordinary linear mapping, and the space of all linear mappings from V to W is denoted L(V;W).
The universal characterization of the tensor product implies that, for each multilinear function
there exists a unique linear function
for all vi ∈ V and αi ∈ V∗.
Using the universal property, it follows that the space of (m,n)-tensors admits a natural isomorphism
In the formula above, the roles of V and V* are reversed. In particular, one has
Differential geometry, physics and engineering must often deal with tensor fields on smooth manifolds. The term tensor is sometimes used as a shorthand for tensor field. A tensor field expresses the concept of a tensor that varies from point to point.
- Abraham, Ralph; Marsden, Jerrold E. (1985), Foundations of Mechanics (2 ed.), Reading, Mass.: Addison-Wesley, ISBN 0-201-40840-6.
- Bourbaki, Nicolas (1989), Elements of mathematics, Algebra I, Springer-Verlag, ISBN 3-540-64243-9.
- de Groote, H. F. (1987), Lectures on the Complexity of Bilinear Problems, Lecture Notes in Computer Science, 245, Springer, ISBN 3-540-17205-X.
- Halmos, Paul (1974), Finite-dimensional Vector Spaces, Springer, ISBN 0-387-90093-4.
- Jeevanjee, Nadir (2011), An Introduction to Tensors and Group Theory for Physicists, ISBN 978-0-8176-4714-8
- Knuth, Donald (1998) , The Art of Computer Programming vol. 2 (3rd ed.), pp. 145–146, ISBN 978-0-201-89684-8.