
Inmathematics, alinear combination orsuperposition is anexpression constructed from aset of terms by multiplying each term by a constant and adding the results (e.g. a linear combination ofx andy would be any expression of the formax +by, wherea andb are constants).[1][2][3] The concept of linear combinations is central tolinear algebra and related fields of mathematics. Most of this article deals with linear combinations in the context of avector space over afield, with some generalizations given at the end of the article.
LetV be avector space over the fieldK. As usual, we call elements ofVvectors and call elements ofKscalars.Ifv1,...,vn are vectors anda1,...,an are scalars, then thelinear combination of those vectors with those scalars as coefficients is
There is some ambiguity in the use of the term "linear combination" as to whether it refers to the expression or to its value. In most cases the value is emphasized, as in the assertion "the set of all linear combinations ofv1,...,vn always forms a subspace". However, one could also say "two different linear combinations can have the same value" in which case the reference is to the expression. The subtle difference between these uses is the essence of the notion oflinear dependence: a familyF of vectors is linearly independent precisely if any linear combination of the vectors inF (as value) is uniquely so (as expression). In any case, even when viewed as expressions, all that matters about a linear combination is the coefficient of eachvi; trivial modifications such as permuting the terms or adding terms with zero coefficient do not produce distinct linear combinations.
In a given situation,K andV may be specified explicitly, or they may be obvious from context. In that case, we often speak ofa linear combination of the vectorsv1,...,vn, with the coefficients unspecified (except that they must belong toK). Or, ifS is asubset ofV, we may speak ofa linear combination of vectors in S, where both the coefficients and the vectors are unspecified, except that the vectors must belong to the setS (and the coefficients must belong toK). Finally, we may speak simply ofa linear combination, where nothing is specified (except that the vectors must belong toV and the coefficients must belong toK); in this case one is probably referring to the expression, since every vector inV is certainly the value of some linear combination.
Note that by definition, a linear combination involves onlyfinitely many vectors (except as described in the§ Generalizations section.However, the setS that the vectors are taken from (if one is mentioned) can still beinfinite; each individual linear combination will only involve finitely many vectors.Also, there is no reason thatn cannot bezero; in that case, we declare by convention that the result of the linear combination is thezero vector inV.
This section includes alist of references,related reading, orexternal links,but its sources remain unclear because it lacksinline citations. Please helpimprove this section byintroducing more precise citations.(August 2013) (Learn how and when to remove this message) |
Let the fieldK be the setR ofreal numbers, and let the vector spaceV be theEuclidean spaceR3.Consider the vectorse1 = (1,0,0),e2 = (0,1,0) ande3 = (0,0,1).Thenanyvector inR3 is a linear combination ofe1,e2, and e3.
To see that this is so, take an arbitrary vector (a1,a2,a3) inR3, and write:
LetK be the setC of allcomplex numbers, and letV be the set CC(R) of allcontinuous functions from thereal lineR to thecomplex planeC.Consider the vectors (functions)f andg defined byf(t) :=eit andg(t) :=e−it.(Here,e is thebase of the natural logarithm, about 2.71828..., andi is theimaginary unit, a square root of −1.)Some linear combinations off andg are:
On the other hand, the constant function 3 isnot a linear combination off andg. To see this, suppose that 3 could be written as a linear combination ofeit ande−it. This means that there would exist complex scalarsa andb such thataeit +be−it = 3 for all real numberst. Settingt = 0 andt = π gives the equationsa +b = 3 anda +b = −3, and clearly this cannot happen.(SeeEuler's identity.)
LetK beR,C, or any field, and letV be the setP of allpolynomials with coefficients taken from the fieldK.Consider the vectors (polynomials)p1 := 1,p2 :=x + 1, andp3 :=x2 +x + 1.
To find out whetherx2 − 1 is a linear combination ofp1,p2, andp3,we consider an arbitrary linear combination of these vectors and try to see when it equals the desired vectorx2 − 1.Picking arbitrary coefficientsa1,a2, anda3, we want
Multiplying the polynomials out, this means
and collecting like powers ofx, we get
Two polynomials are equalif and only if their corresponding coefficients are equal, so we can conclude
Thissystem of linear equations can easily be solved.First, the first equation simply says thata3 is 1.Knowing that, we can solve the second equation fora2, which comes out to −1.Finally, the last equation tells us thata1 is also −1.Therefore, the only possible way to get a linear combination is with these coefficients.Indeed,
sox2 − 1is a linear combination ofp1,p2, and p3.
On the other hand, if we try to makex3 − 1 a linear combination ofp1,p2, andp3, then following the same process as before, we get the equation
However, when we set corresponding coefficients equal in this case, the equation forx3 is
which is always false.Therefore, there is no way for this to work, andx3 − 1 isnot a linear combination ofp1,p2, and p3.
Take an arbitrary fieldK, an arbitrary vector spaceV, and letv1,...,vn be vectors (inV).It is interesting to consider the set ofall linear combinations of these vectors.This set is called thelinear span (or justspan) of the vectors, sayS = {v1, ...,vn}. We write the span ofS as span(S)[4][5] or sp(S):
Suppose that, for some sets of vectorsv1,...,vn,a single vector can be written in two different ways as a linear combination of them:
This is equivalent, by subtracting these (), to saying a non-trivial combination is zero:[6][7]
If that is possible, thenv1,...,vn are calledlinearly dependent; otherwise, they arelinearly independent.Similarly, we can speak of linear dependence or independence of an arbitrary setS of vectors.
IfS is linearly independent and the span ofS equalsV, thenS is abasis forV.
By restricting the coefficients used in linear combinations, one can define the related concepts ofaffine combination,conical combination, andconvex combination, and the associated notions of sets closed under these operations.
| Type of combination | Restrictions on coefficients | Name of set | Model space |
|---|---|---|---|
| Linear combination | no restrictions | Vector subspace | |
| Affine combination | Affine subspace | Affinehyperplane | |
| Conical combination | Convex cone | Quadrant,octant, ororthant | |
| Convex combination | and | Convex set | Simplex |
Because these are morerestricted operations, more subsets will be closed under them, so affine subsets, convex cones, and convex sets aregeneralizations of vector subspaces: a vector subspace is also an affine subspace, a convex cone, and a convex set, but a convex set need not be a vector subspace, affine, or a convex cone.
These concepts often arise when one can take certain linear combinations of objects, but not any: for example,probability distributions are closed under convex combination (they form a convex set), but not conical or affine combinations (or linear), andpositive measures are closed under conical combination but not affine or linear – hence one definessigned measures as the linear closure.
Linear and affine combinations can be defined over any field (or ring), but conical and convex combination require a notion of "positive", and hence can only be defined over anordered field (orordered ring), generally the real numbers.
If one allows onlyscalar multiplication, not addition, one obtains a (not necessarily convex)cone; one often restricts the definition to only allowing multiplication by positive scalars.
All of these concepts are usually defined as subsets of an ambient vector space (except for affine spaces, which are also considered as "vector spaces forgetting the origin"), rather than being axiomatized independently.
More abstractly, in the language ofoperad theory, one can consider vector spaces to bealgebras over the operad (the infinitedirect sum, so only finitely many terms are non-zero; this corresponds to only taking finite sums), which parametrizes linear combinations: the vector for instance corresponds to the linear combination. Similarly, one can consider affine combinations, conical combinations, and convex combinations to correspond to the sub-operads where the terms sum to 1, the terms are all non-negative, or both, respectively. Graphically, these are the infinite affine hyperplane, the infinite hyper-octant, and the infinite simplex. This formalizes what is meant by being or the standard simplex being model spaces, and such observations as that every boundedconvex polytope is the image of a simplex. Here suboperads correspond to more restricted operations and thus more general theories.
From this point of view, we can think of linear combinations as the most general sort of operation on a vector space – saying that a vector space is an algebra over the operad of linear combinations is precisely the statement thatall possible algebraic operations in a vector space are linear combinations.
The basic operations of addition and scalar multiplication, together with the existence of anadditive identity and additive inverses, cannot be combined in any more complicated way than the generic linear combination: the basic operations are agenerating set for the operad of all linear combinations.
Ultimately, this fact lies at the heart of the usefulness of linear combinations in the study of vector spaces.
IfV is atopological vector space, then there may be a way to make sense of certaininfinite linear combinations, using the topology ofV.For example, we might be able to speak ofa1v1 +a2v2 +a3v3 + ⋯, going on forever.Such infinite linear combinations do not always make sense; we call themconvergent when they do.Allowing more linear combinations in this case can also lead to a different concept of span, linear independence, and basis.The articles on the various flavors of topological vector spaces go into more detail about these.
IfK is acommutative ring instead of a field, then everything that has been said above about linear combinations generalizes to this case without change.The only difference is that we call spaces like thisVmodules instead of vector spaces.IfK is anoncommutative ring, then the concept still generalizes, with one caveat:since modules over noncommutative rings come in left and right versions, our linear combinations may also come in either of these versions, whatever is appropriate for the given module.This is simply a matter of doing scalar multiplication on the correct side.
A more complicated twist comes whenV is abimodule over two rings,KL andKR.In that case, the most general linear combination looks like
wherea1,...,an belong toKL,b1,...,bn belong toKR, andv1,…,vn belong toV.