Inphysics, especially inmultilinear algebra andtensor analysis,covariance andcontravariance describe how the quantitative description of certain geometric or physical entities changes with achange of basis.[2] Briefly, a contravariant vector is a list of numbers that transforms oppositely to a change of basis, and a covariant vector is a list of numbers that transforms in the same way. Contravariant vectors are often just calledvectors and covariant vectors are calledcovectors ordual vectors. The termscovariant andcontravariant were introduced byJames Joseph Sylvester in 1851.[3][4]
Curvilinear coordinate systems, such ascylindrical orspherical coordinates, are often used in physical and geometric problems. Associated with any coordinate system is a natural choice of coordinate basis for vectors based at each point of the space, and covariance and contravariance are particularly important for understanding how the coordinate description of a vector changes by passing from one coordinate system to another.Tensors are objects inmultilinear algebra that can have aspects of both covariance and contravariance.
In physics, a vector typically arises as the outcome of a measurement or series of measurements, and is represented as a list (ortuple) of numbers such as
The numbers in the list depend on the choice ofcoordinate system. For instance, if the vector represents position with respect to an observer (position vector), then the coordinate system may be obtained from a system of rigid rods, or reference axes, along which the componentsv1,v2, andv3 are measured. For a vector to represent a geometric object, it must be possible to describe how it looks in any other coordinate system. That is to say, the components of the vectors willtransform in a certain way in passing from one coordinate system to another.
A simple illustrative case is that of aEuclidean vector. For a vector, once a set of basis vectors has been defined, then the components of that vector will always varyopposite to that of the basis vectors. That vector is therefore defined as acontravariant tensor. Take a standard position vector for example. By changing the scale of the reference axes from meters to centimeters (that is,dividing the scale of the reference axes by 100, so that the basis vectors now are meters long), the components of the measuredpositionvector aremultiplied by 100. A vector's components change scaleinversely to changes in scale to the reference axes, and consequently a vector is called acontravariant tensor.
Avector, which is an example of acontravariant tensor, has components that transform inversely to the transformation of the reference axes, (with example transformations includingrotation anddilation).The vector itself does not change under these operations; instead, the components of the vector change in a way that cancels the change in the spatial axes. In other words, if the reference axes were rotated in one direction, the component representation of the vector would rotate in exactly the opposite way. Similarly, if the reference axes were stretched in one direction, the components of the vector, would reduce in an exactly compensating way. Mathematically, if the coordinate system undergoes a transformation described by aninvertible matrixM, so that the basis vectors transform according to, then the components of a vectorv in the original basis ( ) must be similarly transformed via. The components of avector are often represented arranged in a column.
By contrast, acovector has components that transform like the reference axes. It lives in the dual vector space, and represents a linear map from vectors to scalars. The dot product operator involving vectors is a good example of a covector. To illustrate, assume we have a covector defined as , where is a vector. The components of this covector in some arbitrary basis are, with being the basis vectors in the corresponding vector space. (This can be derived by noting that we want to get the correct answer for the dot product operation when multiplying by an arbitrary vector , with components). The covariance of these covector components is then seen by noting that if a transformation described by aninvertible matrixM were to be applied to the basis vectors in the corresponding vector space,, then the components of the covector will transform with the same matrix, namely,. The components of acovector are often represented arranged in a row.
A third concept related to covariance and contravariance isinvariance. Ascalar (also called type-0 or rank-0 tensor) is an object that does not vary with the change in basis. An example of a physicalobservable that is a scalar is themass of a particle. The single, scalar value of mass is independent to changes in basis vectors and consequently is calledinvariant. The magnitude of a vector (such asdistance) is another example of an invariant, because it remains fixed even if geometrical vector components vary. (For example, for a position vector of length meters, if allCartesian basis vectors are changed from meters in length to meters in length, the length of the position vector remains unchanged at meters, although the vector components will all increase by a factor of). The scalar product of a vector and a covector is invariant, because one has components that vary with the base change, and the other has components that vary oppositely, and the two effects cancel out. One thus says that covectors aredual to vectors.
Thus, to summarize:
The general formulation of covariance and contravariance refers to how the components of a coordinate vector transform under achange of basis (passive transformation).[5] Thus letV be avector space of dimensionn over afield ofscalarsS, and let each off = (X1, ...,Xn) andf′ = (Y1, ...,Yn) be abasis ofV.[note 1] Also, let thechange of basis fromf tof′ be given by
1 |
for someinvertiblen×n matrixA with entries.Here, each vectorYj of thef′ basis is a linear combination of the vectorsXi of thef basis, so that
which are the columns of the matrix product.
A vector inV is expressed uniquely as alinear combination of the elements of thef basis as
2 |
wherevi[f] are elements of the fieldS known as thecomponents ofv in thef basis. Denote thecolumn vector of components ofv byv[f]:
so that (2) can be rewritten as a matrix product
The vectorv may also be expressed in terms of thef′ basis, so that
However, since the vectorv itself is invariant under the choice of basis,
The invariance ofv combined with the relationship (1) betweenf andf′ implies that
giving the transformation rule
In terms of components,
where the coefficients are the entries of theinverse matrix ofA.
Because the components of the vectorv transform with theinverse of the matrixA, these components are said totransform contravariantly under a change of basis.
The wayA relates the two pairs is depicted in the following informal diagram using an arrow. The reversal of the arrow indicates a contravariant change:
Alinear functionalα onV is expressed uniquely in terms of itscomponents (elements inS) in thef basis as
These components are the action ofα on the basis vectorsXi of thef basis.
Under the change of basis fromf tof′ (via1), the components transform so that
3 |
Denote therow vector of components ofα byα[f]:
so that (3) can be rewritten as the matrix product
Because the components of the linear functional α transform with the matrixA, these components are said totransform covariantly under a change of basis.
The wayA relates the two pairs is depicted in the following informal diagram using an arrow. A covariant relationship is indicated since the arrows travel in the same direction:
Had a column vector representation been used instead, the transformation law would be thetranspose
The choice of basisf on the vector spaceV defines uniquely a set of coordinate functions onV, by means of
The coordinates onV are therefore contravariant in the sense that
Conversely, a system ofn quantitiesvi that transform like the coordinatesxi onV defines a contravariant vector (or simply vector). A system ofn quantities that transform oppositely to the coordinates is then a covariant vector (or covector).
This formulation of contravariance and covariance is often more natural in applications in which there is a coordinate space (amanifold) on which vectors live astangent vectors orcotangent vectors. Given a local coordinate systemxi on the manifold, the reference axes for the coordinate system are thevector fields
This gives rise to the framef = (X1, ...,Xn) at every point of the coordinate patch.
Ifyi is a different coordinate system and
then the framef' is related to the framef by the inverse of theJacobian matrix of the coordinate transition:
Or, in indices,
A tangent vector is by definition a vector that is a linear combination of the coordinate partials. Thus a tangent vector is defined by
Such a vector is contravariant with respect to change of frame. Under changes in the coordinate system, one has
Therefore, the components of a tangent vector transform via
Accordingly, a system ofn quantitiesvi depending on the coordinates that transform in this way on passing from one coordinate system to another is called a contravariant vector.
In a finite-dimensionalvector spaceV over a fieldK with a non-degenerate symmetricbilinear formg :V ×V →K (which may be referred to as themetric tensor), there is little distinction between covariant and contravariant vectors, because thebilinear form allows covectors to be identified with vectors. That is, a vectorv uniquely determines a covectorα via
for all vectorsw. Conversely, each covectorα determines a unique vectorv by this equation. Because of this identification of vectors with covectors, one may speak of thecovariant components orcontravariant components of a vector, that is, they are just representations of the same vector using thereciprocal basis.
Given a basisf = (X1, ...,Xn) ofV, there is a unique reciprocal basisf# = (Y1, ...,Yn) ofV determined by requiring that
theKronecker delta. In terms of these bases, any vectorv can be written in two ways:
The componentsvi[f] are thecontravariant components of the vectorv in the basisf, and the componentsvi[f] are thecovariant components ofv in the basisf. The terminology is justified because under a change of basis,
where is an invertible matrix, and thematrix transpose has its usual meaning.
In the Euclidean plane, thedot product allows for vectors to be identified with covectors. If is a basis, then the dual basis satisfies
Thus,e1 ande2 are perpendicular to each other, as aree2 ande1, and the lengths ofe1 ande2 normalized againste1 ande2, respectively.
For example,[6] suppose that we are given a basise1,e2 consisting of a pair of vectors making a 45° angle with one another, such thate1 has length 2 ande2 has length 1. Then the dual basis vectors are given as follows:
Applying these rules, we find
and
Thus the change of basis matrix in going from the original basis to the reciprocal basis is
since
For instance, the vector
is a vector with contravariant components
The covariant components are obtained by equating the two expressions for the vectorv:
so
In the three-dimensionalEuclidean space, one can also determine explicitly the dual basis to a given set ofbasis vectorse1,e2,e3 ofE3 that are not necessarily assumed to be orthogonal nor of unit norm. The dual basis vectors are:
Even when theei andei are notorthonormal, they are still mutually reciprocal:
Then the contravariant components of any vectorv can be obtained by thedot product ofv with the dual basis vectors:
Likewise, the covariant components ofv can be obtained from the dot product ofv with basis vectors, viz.
Thenv can be expressed in two (reciprocal) ways, viz.
or
Combining the above relations, we have
and we can convert between the basis and dual basis with
and
If the basis vectors areorthonormal, then they are the same as the dual basis vectors.
The following applies to any vector space of dimensionn equipped with a non-degenerate commutative and distributive dot product, and thus also to the Euclidean spaces of any dimension.
All indices in the formulas run from 1 ton. TheEinstein notation for the implicit summation of the terms with the same upstairs (contravariant) and downstairs (covariant) indices is followed.
The historical and geometrical meaning of the termscontravariant andcovariant will be explained at the end of this section.
Considering this figure for the case of an Euclidean space with, since, if we want to express in terms of the covariant basis, we have to multiply the basis vectors by the coefficients.
With and thus and fixed, if the module of increases, the value of the component decreases, and that's why they're calledcontra-variant (with respect to the variation of the basis vectors module).
Symmetrically, corollary (7) states that the components equal the dot product between the vector and the covariant basis vectors, and since this is directly proportional to the basis vectors module, they're calledco-variant.
If we consider the dual (contravariant) basis, the situation is perfectly specular: the covariant components arecontra-variant with respect to the module of the dual basis vectors, while the contravariant components areco-variant.
So in the end it all boils down to a matter of convention: historically the first non-orthonormal basis of the vector space of choice was called "covariant", its dual basis "contravariant", and the corresponding components named specularly.
If the covariant basis becomesorthonormal, the dual contravariant basis aligns with it and the covariant components collapse into the contravariant ones, the most familiar situation when dealing with geometrical Euclidean vectors. and become the identity matrix, and:
If the metric is non-Euclidean, but for instanceMinkowskian like in thespecial relativity andgeneral relativity theories, the basis are never orthonormal, even in the case of special relativity where and become, for. In this scenario, the covariant and contravariant components always differ.
The distinction between covariance and contravariance is particularly important for computations withtensors, which often havemixed variance. This means that they have both covariant and contravariant components, or both vector and covector components. The valence of a tensor is the number of covariant and contravariant terms, and inEinstein notation, covariant components have lower indices, while contravariant components have upper indices. The duality between covariance and contravariance intervenes whenever a vector or tensor quantity is represented by its components, although moderndifferential geometry uses more sophisticatedindex-free methods to represent tensors.
Intensor analysis, acovariant vector varies more or less reciprocally to a corresponding contravariant vector. Expressions for lengths, areas and volumes of objects in the vector space can then be given in terms of tensors with covariant and contravariant indices. Under simple expansions and contractions of the coordinates, the reciprocity is exact; under affine transformations the components of a vector intermingle on going between covariant and contravariant expression.
On amanifold, atensor field will typically have multiple, upper and lower indices, where Einstein notation is widely used. When the manifold is equipped with ametric, covariant and contravariant indices become very closely related to one another. Contravariant indices can be turned into covariant indices bycontracting with the metric tensor. The reverse is possible by contracting with the (matrix) inverse of the metric tensor. Note that in general, no such relation exists in spaces not endowed with a metric tensor. Furthermore, from a more abstract standpoint, a tensor is simply "there" and its components of either kind are only calculational artifacts whose values depend on the chosen coordinates.
The explanation in geometric terms is that a general tensor will have contravariant indices as well as covariant indices, because it has parts that live in thetangent bundle as well as thecotangent bundle.
A contravariant vector is one which transforms like, where are the coordinates of a particle at itsproper time. A covariant vector is one which transforms like, where is a scalar field.
Incategory theory, there arecovariant functors andcontravariant functors. The assignment of thedual space to a vector space is a standard example of a contravariant functor. Contravariant (resp. covariant) vectors are contravariant (resp. covariant) functors from a-torsor to the fundamental representation of. Similarly, tensors of higher degree are functors with values in other representations of. However, some constructions ofmultilinear algebra are of "mixed" variance, which prevents them from being functors.
Indifferential geometry, the components of a vector relative to a basis of thetangent bundle are covariant if they change with the same linear transformation as a change of basis. They are contravariant if they change by the inverse transformation. This is sometimes a source of confusion for two distinct but related reasons. The first is that vectors whose components are covariant (called covectors or1-forms) actuallypull back under smooth functions, meaning that the operation assigning the space of covectors to a smooth manifold is actually acontravariant functor. Likewise, vectors whose components are contravariantpush forward under smooth mappings, so the operation assigning the space of (contravariant) vectors to a smooth manifold is acovariant functor. Secondly, in the classical approach to differential geometry, it is not bases of the tangent bundle that are the most primitive object, but rather changes in the coordinate system. Vectors with contravariant components transform in the same way as changes in the coordinates (because these actually change oppositely to the induced change of basis). Likewise, vectors with covariant components transform in the opposite way as changes in the coordinates.