# Linear Algebra/Null Spaces

Among the three important vector spaces associated with a matrix of order m x n is the **Null Space**. Null spaces apply to linear transformations.

## Range

editLet T be a linear transformation from an m-dimension vector space X to an n-dimensional vector space Y, and let x_{1}, x_{2}, x_{3}, ..., x_{m} be a basis for X and let y_{1}, y_{2}, y_{3}, ..., y_{n} be a basis for Y, and consider its corresponding n × m matrix,

.

The image of X, T(X), is called the range of T. T(A) is obviously a subspace of Y.

Since any element x within X can be expressed as

,

implying that the range of T is the vector space spanned by the vectors T(x_{i}) which is indicated by the columns of the matrix. By a theorem proven earlier, the dimension of the vector space spanned by those vectors is equal to the maximum number of vectors that are linearly independent. Since the linear dependence of columns in the matrix is the same as the linear dependence of the vectors T(x_{i}), the dimension is equal to the maximum number of columns that are linearly independent, which is equal to the rank. We have the following important conclusion:

The dimension of the range of a linear transformation is equal to the rank of its corresponding matrix.

## Null Space

editFor example, consider the matrix: .

The null space of this matrix consists of the set:

It may not be immediately obvious how we found this set but it can be readily checked that any element of this set indeed gives the zero vector on being multiplied by A. Clearly,

as

.

## Null Space as a vector space

editIt is easy to show that the null space is in fact a vector space. If we identify a n x 1 column matrix with an element of the n dimensional Euclidean space then the null space becomes its subspace with the usual operations. The null space may also be treated as a subspace of the vector space of all n x 1 column matrices with matrix addition and scalar multiplication of a matrix as the two operations.

To show that the null space is indeed a vector space it is sufficient to show that

and

These are true due to the distributive law of matrices. The details of the proof are left to the reader as an exercise.

## Properties

edit### Null spaces of row equivalent matrices

editIf A and B are two row equivalent matrices then they share the same null space. This fact, which is in fact a little theorem, can be proved as follows:

Suppose x is an element of the null space of A. Then Ax = 0. Also since A is row equivalent to B so where each is an elementary matrix. (Recall that an elementary matrix is the matrix obtained from performing any elementary row operation.) Now,

and so x is in the null space of B as well. So the null space of A is contained in that of B. Similarly the null space of B is contained in that of A. It is now clear that A and B have the same null space.

## Basis of Null Space

editAs the null space of a matrix is a vector space, it is natural to wonder what its basis will be. Of course, since the null space is a subspace of , its basis can have at most n elements in it. The number of elements in the basis of the null space is important and is called the **nullity** of A.
To find out the basis of the null space of A we follow the following steps:

- First convert the given matrix into row echelon form say U.
- Next circle the first non zero entries in each row.
- Call the variable as a basic variable if the first column has a circled entry, and call it a free variable if the first column doesn't have a circled entry. Similarly call the variable basic if the second column has a non zero entry and free otherwise. In this way name n variables .
- If for any i, is a free variable, then let be the solution obtained by solving the system Ux = 0 where all the free variables are exactly 0, except for which is 1. If is not a free variable don't do anything.
- Repeat the above step for all the free variables getting vectors etc in the process.
- The set is the required basis.

The key point in the above algorithm was that A and U have the same null space. For a complete proof of why the algorithm works we refer the reader to the excellent text book given in the references by Hoffman and Kunze.

Let us look at an example:

Suppose

The first step involves reducing A to its row echelon form U.

Now

We encircle the first non zero entries in each row by brackets:

Clearly the free variables are and and the rest and are basic variables. Now we shall solve the system Ux = 0 with to get the vector . Thus we need to solve,

This reduces to the following system on matrix multiplication:

It is clear from here that is the solution.

Thus . Similarly is found to be .

The set is the basis of the null space and the nullity of the matrix A is 2. In fact this method gives us a way to describe the null space as well which would be: (Why? - Because the linear combination of solutions is also a solution)

## Implications of nullity being zero

editThe example given above gives no hint as to what happens when there are no free variables in the row echelon form of A. All we said that in step 4 of our algorithm was that if is not a free variable then *don't do anything*. Following that logic, if no variable is free then we keep on doing nothing, leading to the conclusion that *if no variable is free then the basis of the null space is an empty set i.e. *. In that case we say that the nullity of the null space is 0. Note that the null space itself is not empty and contains precisely one element which is the zero vector.

Now suppose that A is any matrix of order m x n with columns . Each is a vector in the m-dimensional space. If the nullity of A is zero, then it follows that Ax=0 has only the zero vector as the solution.

More precisely,

has the trivial solution only. This implies that nullity being zero makes it necessary for the columns of A to be linearly independent. By retracing our steps we can show that the converse is true as well.

Let us examine the special case of a square matrix, i.e. when m = n. Now if the nullity is zero then there is no free variable in the row reduced echelon form of the matrix A, which is say U. Hence each row contains a pivot, or a leading non zero entry. In that case U must be of the form,

or U must precisely be the identity matrix I. Conversely, if A is row equivalent to I then Ax = 0 and Ix = 0 have the same solutions, due to their being equivalent. Since Ix = 0 has only the trivial solution x = 0, so does Ax = 0. It follows that the null space of A is merely {0} and so the nullity of A is 0.

Thus nullity of A is 0 A is row equivalent to I.

Now if A is row equivalent to I then where each is an elementary matrix. Since a product of invertible matrices is invertible and each is invertible so A is invertible. Conversely if A was invertible, and U its row reduced echelon form then which is clearly invertible (by virtue of being a product of invertible matrices). Now a matrix containing a zero row can never be invertible (why?), so U has pivots in each row. It follows that there are n pivots all equal to 1, with zeros above and below them and so U = I. Thus A is row equivalent to I.

In summary, A is row equivalent to I A is invertible.

We can collect the entire argument in this section, to state the:

**Theorem**: For a square matrix of order n, the following are equivalent:

- A is invertible.
- Nullity of A is 0.
- A is row equivalent to the identity matrix.
- Columns of A are linearly independent.
- The system Ax = 0 has only the trivial solution.
- A is a product of elementary matrices.

It will be a good exercise for the reader at this stage to try to rewrite the proof of the theorem in detail.

## Exercises

edit- Evaluate null spaces and bases for:
- Show that null space of a matrix is a vector space.
- Prove the theorem regarding invertibility of a square matrix. Also by showing that A is invertible iff A is, show that the condition that the rows are linearly independent can be added to the list.
- Is the solution set for Ax = b where b is a non zero vector (i.e. has at least one component non zero) a vector space? Give reasons.
- Let r be the number of basic variables associated with a n order matrix A (which is equal to those associated with its row echelon form). Show that A is invertible if and only if r = n.