Intermediate treatment of tensors

From Wikipedia, the free encyclopedia

 Note
The following is a modern component-based treatment of tensors (sometimes called the "classical treatment" of tensors). Read the article tensor for a simple description of tensors, or see the component-free treatment of tensors for a more abstract treatment. For an even more traditional approach, see classical treatment of tensors.
Note that the word "tensor" is often used as a shorthand for "tensor field", a concept which defines a tensor value at every point in a manifold. To understand tensor fields, you need to first understand tensors.

A tensor is the mathematical idealization of a geometric or physical quantity whose analytic description, relative to a fixed frame of reference, consists of an array of numbers.

This way of viewing tensors, called tensor analysis, was used by Einstein and is generally preferred by physicists. It is, very grossly, a generalization of the concept of vectors and matrices and allows the writing of equations independently of any given coordinate system.

Contents

[edit] Overview

It should be noted that the array-of-numbers representation of a tensor is not the same thing as the tensor. An image and the object represented by the image are not the same thing. The mass of a stone is not a number. Rather, the mass can be described by a number relative to some specified unit mass. Similarly, a given numerical representation of a tensor only makes sense in a particular coordinate system.

Some well known examples of tensors in geometry are quadratic forms, and the curvature tensor. Examples of physical tensors are the energy-momentum tensor and the polarization tensor.

Geometric and physical quantities may be categorized by considering the degrees of freedom inherent in their description. The scalar quantities are those that can be represented by a single number—speed, mass, temperature, for example. There are also vector-like quantities such as force that require a list of numbers for their description (so that direction can be accounted for). Finally, quantities such as quadratic forms naturally require a multiply-indexed array for their representation. These latter quantities can only be conceived of as tensors.

Actually, the tensor notion is quite general and applies to all of the above examples; scalars and vectors are special kinds of tensors. The feature that distinguishes a scalar from a vector, and distinguishes both of those from a more general tensor quantity is the number of indices in the representing array. This number is called the rank (or the order) of a tensor. Thus, scalars are rank zero tensors (with no indices at all) and vectors are rank one tensors.

It is also necessary to distinguish between two types of indices, depending on whether the corresponding numbers transform covariantly or contravariantly relative to a change in the frame of reference. Contravariant indices are written as superscripts, while the covariant indices are written as subscripts. The type (or valence) of a tensor is the pair (p,q), where p is the number of contravariant and q the number of covariant indices, respectively. Note that a tensor of type (p,q) has a rank of p + q. It is customary to represent the actual tensor, as a standalone entity, by a bold-face symbol such as \mathbf{T}. The corresponding array of numbers for a type (p,q) tensor is denoted by the symbol T^{i_1\ldots i_p}_{j_1\ldots j_q}, where the superscripts and subscripts are indices that vary from 1 to n. The number n, the range of the indices, is called the dimension of the tensor; the total number of degrees of freedom required for the specification of a particular tensor is the dimension of the tensor raised to the power of the tensor's rank.

Again, it must be emphasized that the tensor \mathbf{T} and the representing array T^{i_1\ldots i_q}_{j_1\ldots j_p} are not the same thing. The values of the representing array are given relative to some frame of reference, and undergo a linear transformation when the frame is changed.

Finally, it must be mentioned that most physical and geometric applications are concerned with tensor fields, that is to say tensor valued functions, rather than tensors themselves. Some care is required, because it is common to see a tensor field called simply a tensor. There is a difference, however; the entries of a tensor array T^{i_1\ldots i_q}_{j_1\ldots j_p} are numbers, whereas the entries of a tensor field are functions. The present entry treats the purely algebraic aspect of tensors. Tensor field concepts, which typically involve derivatives of some kind, are discussed elsewhere.

[edit] Definition

The formal definition of a tensor quantity begins with a finite-dimensional vector space \mathcal{U}, which furnishes the uniform "building blocks" for tensors of all valences. In typical applications, \mathcal{U} is the tangent space at a point of a manifold; the elements of \mathcal{U} typically represent physical quantities such as velocities or forces. The space of (p,q)-valent tensors, denoted here by \mathcal{U}^{p,q} is obtained by taking the tensor product of p copies of \mathcal{U} and q copies of the dual vector space \mathcal{U}^*. To wit,

\mathcal{U}^{p,q} =  \left\{\mathcal{U}\otimes\ldots\otimes\mathcal{U}\right\} \otimes \left\{\mathcal{U}^*\otimes\ldots\otimes\mathcal{U}^*\right\}

In order to represent a tensor by a concrete array of numbers, we require a frame of reference, which is essentially a basis of \mathcal{U}, say \mathbf{e}_1,\ldots,\mathbf{e}_n \in \mathcal{U}. Every vector in \mathcal{U} can be "measured" relative to this basis, meaning that for every \mathbf{v}\in\mathcal{U} there exist unique scalars vi, such that (note the use of the Einstein notation)

\mathbf{v} = v^i\mathbf{e}_i

These scalars are called the components of \mathbf{v} relative to the frame in question.

Let \varepsilon^1,\ldots,\varepsilon^n\in\mathcal{U}^* be the corresponding dual basis, i.e.,

\varepsilon^i(\mathbf{e}_j) = \delta^i {}_j,

where the latter is the Kronecker delta array. For every covector \mathbf{\alpha}\in\mathcal{U}^* there exists a unique array of components αi such that

\mathbf{\alpha} = \alpha_i\, \varepsilon^i.

More generally, every tensor \mathbf{T}\in\mathcal{U}^{p,q} has a unique representation in terms of components. That is to say, there exists a unique array of scalars T^{i_1\ldots i_p}_{j_1\ldots j_q} such that

\mathbf{T} = T^{i_1\ldots i_p}_{j_1\ldots j_q}\, \mathbf{e}_{i_1} \otimes \ldots\otimes \mathbf{e}_{i_q} \otimes \varepsilon^{j_1}\otimes\ldots\otimes \varepsilon^{j_p}.

[edit] Transformation rules

Next, suppose that a change is made to a different frame of reference, say \hat{\mathbf{e}}_1,\ldots,\hat{\mathbf{e}}_n\in\mathcal{U}. Any two frames are uniquely related by an invertible transition matrix Aij, having the property that for all values of j we have the frame transformation rule

\hat{\mathbf{e}}_j = A^i {}_j\, \mathbf{e}_i.

Let \mathbf{v}\in\mathcal{U} be a vector, and let vi and \hat{v}^i denote the corresponding component arrays relative to the two frames. From

\mathbf{v} = v^i\mathbf{e}_i = \hat{v}^i\hat{\mathbf{e}}_i,

and from the frame transformation rule we infer the vector transformation rule

\hat{v}^i = B^i {}_j\, v^j,

where Bij is the matrix inverse of Aij, i.e.,

AikBkj = δij.

Thus, the transformation rule for a vector's components is contravariant to the transformation rule for the frame of reference. It is for this reason that the superscript indices of a vector are called contravariant.

To establish the transformation rule for covectors, we note that the transformation rule for the dual basis takes the form

\hat{v}e^i = B^i {}_j \, \varepsilon^j,

and that

v^i = \varepsilon^i(\mathbf{v}),

while

\hat{v}^i = \hat{v}e^i(\mathbf{v}).

The transformation rule for covector components is covariant. Let \mathbf{\alpha}\in \mathcal{U}^* be a given covector, and let αi and \hat{\alpha}_i be the corresponding component arrays. Then

\hat{\alpha}_j = A^i {}_j \alpha_i.

The above relation is easily established. We need only remark that

\alpha_i = \mathbf{\alpha}(\mathbf{e}_i),

and that

\hat{\alpha}_j = \mathbf{\alpha}(\hat{\mathbf{e}}_j),

and then use the transformation rule for the frame of reference.

In light of the above discussion, we see that the transformation rule for a general type (p,q) tensor takes the form

\hat{T}^{i_1\ldots i_q}_{\,j_1\ldots j_p} = A^{i_1} {}_{k_1}\cdots A^{i_q} {}_{k_q} B^{l_1} {}_{j_1}\cdots B^{l_1} {}_{j_p} T^{k_1\ldots k_q}_{l_1\ldots l_p}.

[edit] See also

[edit] Further reading

  • Bernard Schutz, Geometrical methods of mathematical physics, Cambridge University Press, 1980.
  • Schaum's Outline of Tensor Calculus
  • Synge and Schild, Tensor Calculus, Toronto Press: Toronto, 1949
  • James Munkres, Analysis on Manifolds, Westview Press, 1991. Chapter six gives a "from scratch" introduction to covariant tensors.

An earlier version of this article was adapted from the GFDL article on tensors at http://planetmath.org/encyclopedia/Tensor.html from PlanetMath, written by Robert Milson and others