Advanced Mathematics for Engineers and Scientists/Vector Spaces: Mathematic Playgrounds

Vector Spaces: Mathematic Playgrounds edit

The study of partial differential equations requires a clear definition of what kind of numbers are being dealt with and in what way. PDEs are normally studied in certain kinds of vector spaces, which have a number of properties and rules associated with them which make possible the analysis and unifies many notions.

The Real Field edit

A field is a set that is bundled with two operations on the set called addition and multiplication which obey certain rules, called axioms. The letter   will be used to represent the field, and from definition a field requires the following (  and   are in  ):

  • Closure under addition and multiplication: the addition and multiplication of field members produces members of the same field.
  • Addition and multiplication are associative:   and  .
  • Addition and multiplication are commutative:   and  .
  • Addition and multiplication are distributive:   and  .
  • Existence of additive identity: there is an element in   notated 0, sometimes called the sum of no numbers, such that  .
  • Existence of multiplicative identity: there is an element in   notated 1 different from 0, sometimes called the product of no numbers, such that  .
  • Existence of additive inverse: there is an element in   associated with   notated   such that  .
  • Existence of multiplicative inverse: there is an element in   associated with   (if   is nonzero), notated   such that  .

These are called the field axioms. The field that we deal with, by far the most common one, is the real field. The set associated with the real field is the set of real numbers, and addition and multiplication are the familiar operations that everyone knows about.

Another example of a set that can form a field is the set of rational numbers, numbers which are expressible as the ratio of two integers. An example of a common set that doesn't form a field is the set of integers: there generally is no multiplicative inverse since the reciprocal of an integer generally is not an integer.

Note that when we say that an object is in  , what is meant is that the object is a member of the set associated in the field and that it complies with the field axioms.

The Vector edit

Most non-mathematics students are taught that vectors are ordered groups ("tuples") of quantities. This is not complete, vectors are a lot more general than that. Informally, a vector is defined as an object that can be scaled and added with other vectors. This will be made more specific soon.

Examples of vectors:

  • The real numbers.
  • Pairs, triples, etc of real numbers.
  • Polynomials.
  • Most functions.

Examples of objects that are not vectors:

  • Members of the extended real numbers. Specifically, the infinity and negative infinity elements neither scale nor add.
  • The integers, at least when scaled by real numbers (since the result will not necessarily be an integer).

An interesting (read: confusing) fact to note is that, by the definition above, matrices and even tensors qualify as vectors since they can be scaled or added, even though these objects are considered generalizations of more "conventional" vectors, and calling a tensor a vector will lead to confusion.

The Vector Space edit

A vector space can be thought of as a generalization of a field.

Letting   represent some field, a vector space   over   is a set of vectors bundled with two operations called vector addition and scalar multiplication, notated:

  • Vector addition:  , where  .
  • Scalar multiplication:  , where   and  .

The members of   are called vectors, and the members of the field   associated with   are called scalars. Note that these operations imply closure (see the first field axiom), so that it does not have to be explicitly stated. Note also that this is essentially where a vector is defined: objects that can be added and scaled. The vector space must comply with the following axioms (  and   are in  ;   and   are in  ):

  • Addition is associative:  .
  • Addition is commutative:  .
  • Scalar multiplication is distributive over vector addition:  .
  • Scalar multiplication is distributive over field addition:  .
  • Scalar and field multiplication are compatible:  .
  • Existence of additive identity: there is an element in   notated 0 such that  .
  • Existence of additive inverse: there is an element in   associated with   notated   such that  .
  • Existence of multiplicative identity: there is an element in   notated 1 different from 0 such that  .

An example of a vector space is one where polynomials are vectors over the real field. An example of a space that is not a vector space is one where vectors are rational numbers over the real field, since scalar multiplication can lead to vectors that are not rational (implied closure under scalar multiplication is violated).

By analogy with linear functions, vectors are linear by nature, hence a vector space is also called a linear space. The name "linear vector space" is also used, but this is somewhat redundant since there is no such thing as a nonlinear vector space. It's now worth mentioning an important quantity called a linear combination (not part of the definition of a vector space, but important):

 

where   is a sequence of field members and   is a sequence of vectors. The fact that a vector can be formed by a linear combination of other vectors is much of the essence of the vector field.

Note that a field over itself qualifies as a vector space. Fields of real numbers and other familiar objects are sometimes called spaces, since distance and other useful concepts apply.

The definition of a vector space is quite general. Note that, for example, there is no mention of any kind of product between vectors, nor is there a notion of the "length" of a vector. The vector space as defined above is very primitive, but it's a starting point: through various extensions, specific vector spaces can have a lot of nice properties and other features that make our playgrounds fun and comfortable. We'll discuss bases (plural of basis) and then take on some specific vector spaces.

The Basis edit

A nonempty subset   of   is called a linear subspace of   if   is itself a vector space. The requirement that   be a vector space can be safely made specific by saying that   is closed under vector addition and scalar multiplication, since the rest of the vector space properties are inherited.

The linear span of a set of   vectors   in   may then be defined as:

 

Where  . The span is the intersection over all choices of  . This concept may be extended so that   is not necessarily finite. The span of   is the intersection of all of the linear subspaces of  .

Now, think of what happens if a vector is removed from the set  . Does the span change? Not necessarily, it may be possible that the remaining vectors in the span are sufficient to "fill in" for the missing vector through linear combination of the remaining vectors.

Let   be a subset of  . If the span of   is the same as the span of  , and if removing a vector from   necessarily changes its span, then the set of vectors   is called a basis of  , and the vectors of   are called linearly independent. It can be proven that a basis can be constructed for every vector space.

Note that the basis is not unique. This obscure definition of a basis is convenient because it is very broad, it is worth understanding fully. An important property of a vector space is that it necessarily has a basis, and that any vector in the space may be written in terms of a linear combination of the members of the basis.

A more understandable (though less elemental) explanation is provided: for a vector space   over the field  , the vectors   form a basis of   (where   are in   and the following   are in  ), satisfying the following properties:

  • The basis vectors are linearly independent: if
 
then
 
without exception.
  • The basis vectors span  : for some given   in  , it is possible to choose   so that
 

The basis vectors of a vector space are usually notated with as  .

Euclidean n-Space edit

As most students are familiar with Euclidean n-space, this section serves more of an example than anything else.

Let   be the field of real numbers, then the vector space   over   is defined to be the space of n-tuples of members of  . In other words, more clearly:

If  , then  , where   are the vectors of the vector space  .

These vectors are called n-dimensional coordinates, and the vector space   is called the real coordinate space; note that coordinates (unlike more general vectors) are often notated in boldface, or else with an arrow over the letter. They are also called spatial vectors, geometric vectors, just "vectors" if the context allows, and sometimes "points" as well, though some authors refuse to consider points as vectors, attributing a "fixed" sense to points so that points can't be added, scaled, or otherwise messed with. Part of the reason for this is that it allows one to say that some vector space is bound to a point, the point being called the origin.

The Euclidean n-space   is the special real coordinate n-space   with some additional structure defined which (finally) gives rise to the geometric notion of (specifically these) vectors.

To begin with, an inner product is first defined, notated with either angle braces or a dot:

 

This quantity, which turns two vectors into a scalar (a member of  ) doesn't have a great deal of geometric meaning until some more structure is defined. In a coordinate space, the dot notation is favored, and this product is often called the "dot product", especially when   or  . The definition of this inner product qualifies   as an inner product space.

Next comes the norm, in terms of the inner product:

 

The notation involving single pipes around the letter x is common, again, when   or  , due to analogy with absolute value, for real and especially complex numbers. For a coordinate space, the norm is often called the length of  . This quickly leads to the notion of the distance between two vectors:

 

Which is simply the length of the vector "from"   to  .

Finally, the angle   between   and   is defined through, for  ,

 

The motivation for this definition of angle, valid for any  , comes from the fact that one can prove that the literal measurable angle between two vectors in   satisfies the above (the norm is motivated similarly). Discussing these 2D angles and distances of course mandates making precise the notion of a vector as an "arrow" (ie, correlating vectors to things you can draw on a sheet of paper), but that would get involved and most are already subconsciously familiar with this and it's not the point of this introduction.

This completes the definition of  . A thorough introduction to Euclidean space isn't very fitting in a text on Partial Differential equations, it is included so that one can see how a familiar vector space can be constructed ground-up through extensions called "structure".

Banach Spaces edit

Banach spaces are more general than Euclidean space, and they begin our departure from vectors as geometric objects into vectors as toys in the crazy world of functional analysis.

To be terse, a Banach space is defined as any complete normed vector space. The details follow.

The Inner Product edit

The inner product is a vector operation which results in a scalar. The vectors are members of a vector space  , and the scalar is a member of the field   associated with  . A vector space on which an inner product is defined is said to be "equipped" with an inner product, and the space is an inner product space. The inner product of   and   is usually notated  .

A truly general definition of the inner product would be long. Normally, if the vectors are real or complex in nature (eg, complex coordinates or real valued functions), the inner product must satisfy the following axioms:

  • Distributive in the first variable:  .
  • Associative in the first variable:  .
  • Nondegeneracy and nonnegativity:  , equality will hold only when  .
  • Conjugate symmetry:  .

Note that if the space is real, the last requirement (the overbar indicates complex conjugation) simplifies to  , and then the first two axioms extend to the second variable.

A desirable property of an inner product is some kind of orthogonality. Two nonzero vectors are said to be orthogonal only if their inner product is zero. Remember that we're talking about vectors in general, not specifically Euclidean.

Inner products are by no means unique, good definitions are what add quality to specific spaces. The Euclidean inner product, for example, defines the Euclidean distance and angle, quantities which form the foundation of Euclidean geometry.

The Norm edit

The norm is usually, though not universally, defined in terms of the inner product, which is why the inner product was discussed first (to be technically correct, a Banach space doesn't necessarily need to have an inner product). The norm is an operation (notated with double pipes) which takes one vector and generates one scalar, necessarily satisfying the following axioms:

  • Scalability:  .
  • The triangle inequality:  .
  • Nonnegativity:  , equality only when  .

The fact that   can be proven from the first two statements above.

Definition requires that   only when   (compare this to the inner product, which can be zero even if fed nonzero vectors); if this condition is relaxed so that   is possible for nonzero vectors, the resulting operation is called a seminorm.

The distance between two vectors   and   is a useful quantity which is defined in terms of the norm:

 

The distance is often called the metric, and a vector space equipped with a distance is called a metric space.

Completeness edit

 
A Cauchy sequence shown in blue.
 
A sequence that is not Cauchy. The elements of the sequence fail to get close to each other as the sequence progresses.

As stated before, a Banach space is defined as a complete normed vector space. The norm was described above, so that all that is left to establish the definition of a Banach space is completeness.

Consider a sequence of vectors   in a vector space  . This sequence of vectors is called a Cauchy sequence if these vectors "tend" toward some "destination" vector, as shown in the pictures at right. Stated precisely, a sequence is a Cauchy sequence if it is always possible to make the distance   arbitrarily small by picking larger values of   and  .

The limit   of a Cauchy sequence is:

 

A vector space   is called complete if every Cauchy sequence has a limit that is also in  . A Banach space is, finally, a vector space equipped with a norm that is complete. Note that completeness implies existence of distance, which means that every Banach space is a metric space.

An example of a vector space that is complete is Euclidean n-space. An example of a vector space that isn't complete is the space of rational numbers over rational numbers: it is possible to form a sequence of rational numbers which limit to an irrational number.

Hilbert spaces edit

Note that the inner product was defined above but not subsequently used in the definition of a Banach space. Indeed, a Banach space must have a norm but doesn't necessarily need to have an inner product. However, if the norm in a Banach space is defined through the inner product by

 

then the resulting special Banach space is called a Hilbert space. Hilbert spaces are important in the study of partial differential equations (some relevance finally!) because many theorems and important results are valid only in Hilbert spaces.