Vector space: Difference between revisions
imported>Paul Wormer |
imported>Paul Wormer |
||
Line 100: | Line 100: | ||
</math> | </math> | ||
Add to the left- and right-hand side of this expression <font style="vertical-align: 10%"><math>\vec{0} = 0\sdot\vec{u}_{q+1} + \cdots + 0\sdot \vec{u}_p</math></font> and we get a contradiction. | Add to the left- and right-hand side of this expression <font style="vertical-align: 10%"><math>\vec{0} = 0\sdot\vec{u}_{q+1} + \cdots + 0\sdot \vec{u}_p</math></font> and we get a contradiction. | ||
===Dimension=== | |||
In general there are infinitely many linearly independent vectors in a vector space. When the ''maximum number'' of linearly independent vectors is finite, say ''n'', the vector space is called of finite dimension ''n''. Otherwise the space is called infinite-dimensional. If ''V''′ is a linear subspace of the ''n''-dimensional space ''V'' (all elements of ''V''′ belong simultaneously to ''V'' ), and ''V''′ contains a set ''B'' of ''m'' linearly independent vectors then ''m'' < ''n'', because ''B'' belongs to the ''n''-dimensional space ''V''. It follows that ''m'' is finite and that all subspaces of finite-dimensional spaces are finite-dimensional. If ''m'' is the maximum number of linearly independent vectors in ''V''′ then this subspace is of dimension ''m'' < ''n''. For finite ''n'' it can be shown that ''V''′ coincides with ''V'' (is an "improper" subspace) if and only if ''n'' = ''m''. | |||
==Examples of vector spaces== | ==Examples of vector spaces== |
Revision as of 08:17, 5 November 2008
Vector spaces, also known as linear spaces, are an abstract mathematical construct with many important applications in the natural sciences, in particular in physics and numerous areas of mathematics. Some vector spaces make sense somewhat intuitively, such as the space of 3D vectors in standard Euclidean space, and the language that we use when talking about these intuitive spaces has been taken to describe the more abstract notion as well. For example, we know how to add vectors and multiply them by real numbers (scalars) in , and these notions of vector addition and scalar multiplication are defined in a more general sense (as we will see below).
Vector spaces are important because many different mathematical objects that at first glance seem unrelated in fact share a common structure. By defining this structure and proving things about it in general, we are then able to apply these results to each specific case without having to re-prove them each time. Besides vectors in that are relatively easy to visualize, we can make a vector space out of for any natural number ; or the complex plane or powers of it, ; or polynomials of degree .
No matter what vector space you have to work with though, it is often useful to keep a picture of either 2D or 3D space in mind. This helps when thinking of things such as orthogonal polynomials or matrices.
Definition
A vector space over a field is a set that satisfies certain axioms (see below) and which is equipped with two operations, vector addition and scalar multiplication. Vector addition is defined as a map
that takes the ordered pair to the vector . Here represents the Cartesian product between sets. Scalar multiplication is defined in a similar way, as a map
that takes the ordered pair to the vector . Note that frequently the dot representing scalar multiplication is omitted, the result being written simply as instead. This is especially common when an inner product will also be defined on the vector space, with the dot then representing the inner product between two vectors. It is important to keep in mind the distinction between scalar multiplication, which multiplies one vector by a scalar, and an inner or scalar product, that combined two vectors to yield a scalar.
Axioms of a vector space
Let be a set, , , and elements of that set, and and scalar elements of a field, . Then is a vector space if the following axioms hold true for all choices of
- 1. is closed under addition
- The vector is also an element of . This is automatically satisfied when the addition operation is defined as being injective as it was above. Care must be taken however if is a subset of some larger set and , as is often the case when looking at subspaces.
- 2. Addition is commutative
- The order in which two vectors are added does not affect the result, .
- 3. Addition is associative
- . This means that even though addition is strictly defined as a binary operation, the object is well defined.
- 4. An additive identity exists in
- Labeled , the additive identity or zero vector satisfies .
- 5. The additive inverse exists in
- A vector can be found such that .
- 6. is closed under scalar multiplication
- The vector is itself an element of .
- 7. Scalar multiplication is distributive over addition in
- . It is important to note that the addition occurring on the left-hand side of this equality is a 'different operation' from the addition on the right-hand side. While the latter is vector addition as defined above, the former is the addition operation defined on the field .
- 8. Vector addition is distributive over scalar multiplication
- . In this case vector addition takes place on both sides of the equality.
- 9. Scalar multiplication is associative
- . This means that the algebraic structure of the underlying field is preserved. Note that the left-hand side of this equality contains two subsequent applications of the scalar multiplication defined above, while the right-hand side contains one scalar multiplication as defined in (that of ), followed by scalar multiplication with the vector .
- 10. The multiplicative identity of provides a scalar multiplicative identity
- , where is the multiplicative identity of the field .
Properties 1 - 5 state that a vector space is an Abelian group with addition as group operation.
These axioms can be expressed concisely in mathematical notation as follows:
Some important theorems
Linear dependence
A system of p ( ≥ 1 ) vectors of a vector space V is called linearly dependent if there exist coefficients (elements in F ) a1, ..., ap not all zero, such that the linear combination is the zero vector in V,
Otherwise, the vectors are called linearly independent. A single vector not equal to the zero vector is obviously linearly independent.
If all a1, ..., ap are zero (in F ) then
If the set is linearly independent then the relation
implies that all a1, ..., ap are zero. Hence a set of p vectors in V is linearly independent if
Every set of vectors containing the zero vector is linearly dependent.
A system of linearly independent vectors remains linearly independent if some vectors are omitted from the system. For, let a subset of the first q vectors , with q < p, be linearly dependent then one or more coefficients not equal to zero can be found while the following is true
Add to the left- and right-hand side of this expression and we get a contradiction.
Dimension
In general there are infinitely many linearly independent vectors in a vector space. When the maximum number of linearly independent vectors is finite, say n, the vector space is called of finite dimension n. Otherwise the space is called infinite-dimensional. If V′ is a linear subspace of the n-dimensional space V (all elements of V′ belong simultaneously to V ), and V′ contains a set B of m linearly independent vectors then m < n, because B belongs to the n-dimensional space V. It follows that m is finite and that all subspaces of finite-dimensional spaces are finite-dimensional. If m is the maximum number of linearly independent vectors in V′ then this subspace is of dimension m < n. For finite n it can be shown that V′ coincides with V (is an "improper" subspace) if and only if n = m.