# Real Analysis/Differentiation in Rn

 Real AnalysisDifferentiation in Rn

## Contents

We will first revise some important concepts of Linear Algebra that are of importance in Multivariate Analysis. The reader with no background in Linear Algebra is advised to refer the book Linear Algebra.

## Vector SpaceEdit

A set ${\displaystyle {\mathcal {V}}}$  is said to be a Vector Space over a field ${\displaystyle F}$  if and only if operations addition and scalar multiplication are defined over it so as to satisfy for all ${\displaystyle \mathbf {v_{1}} ,\mathbf {v_{2}} ,\ldots \in {\mathcal {V}}}$  and ${\displaystyle c_{1},c_{2}\in F}$

(i)Commutativity:${\displaystyle \mathbf {v_{1}} +\mathbf {v_{2}} =\mathbf {v_{2}} +\mathbf {v_{1}} }$

(ii)Associativity:${\displaystyle (\mathbf {v_{1}} +\mathbf {v_{2}} )+\mathbf {v_{3}} =\mathbf {v_{1}} +(\mathbf {v_{2}} +\mathbf {v_{3}} )}$

(iii)Identity:There exists ${\displaystyle \mathbf {0} \in {\mathcal {V}}}$  such that ${\displaystyle \mathbf {v_{1}} +\mathbf {0} =\mathbf {v_{1}} =\mathbf {0} +\mathbf {v_{1}} }$

(iv)Inverse:There exists ${\displaystyle -\mathbf {v_{1}} \in {\mathcal {V}}}$  such that ${\displaystyle \mathbf {v_{1}} +(-\mathbf {v_{1}} )=\mathbf {0} }$

(v):${\displaystyle c\mathbf {v_{1}} +c\mathbf {v_{2}} =c(\mathbf {v_{1}} +\mathbf {v_{2}} )}$

(vi)${\displaystyle c_{1}\mathbf {v} +c_{2}\mathbf {v} =(c_{1}+c_{2})\mathbf {v} }$

(vii)${\displaystyle c_{1}(c_{2}\mathbf {v_{1}} )=c_{2}(c_{1}\mathbf {v_{1}} )=(c_{1}c_{2})\mathbf {v_{1}} }$

Members of a vector space are called "Vectors" and those of the field are called "Scalars". ${\displaystyle \mathbb {R} ^{n}}$ , the set of all polynomials etc. are examples of vector spaces

A set of linearly independant vectors that spans the vector space is said to be a Basis for the vector space.

## Linear TransformationsEdit

Let ${\displaystyle X,Y}$  be vector spaces.

Let ${\displaystyle T:X\to Y}$

We say that ${\displaystyle T}$  is a Linear transformation if and only if for all ${\displaystyle \mathbf {v_{1}} ,\mathbf {v_{2}} \in X}$ ,

(i)${\displaystyle T(\mathbf {v_{1}} +\mathbf {v_{2}} )=T(\mathbf {v_{1}} )+T(\mathbf {v_{2}} )}$

(ii)${\displaystyle T(c\mathbf {v_{1}} )=cT(\mathbf {v_{1}} )}$

As we will see, there are two major ways to define a 'derivative' of a multivariable function. We first present the seemingly more straightforward way of using "Partial Derivatives".

## Directional and Partial DerivativesEdit

Let ${\displaystyle \mathbf {f} :\mathbb {R} ^{n}\to \mathbb {R} ^{m}}$

Let ${\displaystyle \mathbf {a} ,\mathbf {y} \in \mathbb {R} ^{n}}$

We say that ${\displaystyle \mathbf {f} }$  is differentiable at ${\displaystyle \mathbf {a} \in \mathbb {R} ^{n}}$  with respect to vector ${\displaystyle \mathbf {y} }$  if and only if there exists ${\displaystyle \mathbf {L} \in \mathbb {R} ^{m}}$  that satisfies

${\displaystyle \lim _{h\to 0}{\frac {\mathbf {f} (\mathbf {a} +h\mathbf {y} )-\mathbf {f} (\mathbf {a} )}{h}}=\mathbf {L} }$

${\displaystyle \mathbf {L} }$  is said to be the derivative of ${\displaystyle \mathbf {f} }$  at ${\displaystyle \mathbf {a} }$  with respect to ${\displaystyle \mathbf {y} }$  and is written as ${\displaystyle \mathbf {f} '(\mathbf {a} ;\mathbf {y} )}$

When ${\displaystyle \mathbf {y} }$  is a unit vector, the derivative is said to be a partial derivative. Here we will explicitly define partial derivatives and see some of their properties.

Let ${\displaystyle f}$  be a real multivariate function defined on an open subset ${\displaystyle \Omega }$  of ${\displaystyle \mathbb {R} ^{n}}$

${\displaystyle f:\Omega \longrightarrow \mathbb {R} }$ .

Then the partial derivative at some parameter ${\displaystyle (x_{1},...,x_{n})}$  with respect to the coordinate ${\displaystyle x_{i}}$  is defined as the following limit

${\displaystyle \lim _{h\rightarrow 0}{f(x_{1},\ldots ,x_{i}+h,\ldots ,x_{n})-f(x_{1},\ldots ,x_{i},\ldots ,x_{n}) \over h}={\partial f \over \partial x_{i}}}$ .

${\displaystyle f}$  is said to be differentiable at this parameter ${\displaystyle (x_{1},...,x_{n})}$  if the difference ${\displaystyle f(x_{1},...,x_{i}+h,...,x_{n})-f(x_{1},...,x_{i},...,x_{n})}$  is equivalent up to first order in h to a linear form L (of h), that is

${\displaystyle f(x_{1},...,x_{i}+h,...,x_{n})-f(x_{1},...,x_{i},...,x_{n})=L\times h+o(\|h\|).}$

The linear form L is then said to be the differential of ${\displaystyle f}$  at ${\displaystyle (x_{1},...,x_{n})}$ , and is written as ${\displaystyle Df|_{(x_{1},\ldots ,x_{n})}}$  or sometimes ${\displaystyle \mathrm {d} f(x_{1},\ldots ,x_{n})}$ .

In this case, where ${\displaystyle f}$  is differentiable at ${\displaystyle (x_{1},\ldots ,x_{n})}$ , by linearity we can write

${\displaystyle \mathrm {d} f={\partial f \over \partial x_{1}}\mathrm {d} x_{1}+\ldots +{\partial f \over \partial x_{n}}\mathrm {d} x_{n}}$

${\displaystyle f}$  is said to be continuously differentiable if its differential is defined at any parameter in its domain, and if the differential is varying continuously relative to the parameter ${\displaystyle (x_{1},...,x_{n})}$ , that is if it coordinates (as a linear form) ${\displaystyle \partial f \over \partial x_{1}}$  are varying continuously.

In case partial derivatives exists but ${\displaystyle f}$  is not differentiable, and sometimes not even continuous exempli gratia

${\displaystyle f:(x,y)\mapsto {(xy)^{2} \over (x^{2}+y^{2})}}$

(and ${\displaystyle f(0,0)=0}$ ) we say that ${\displaystyle f}$  is separably differentiable.

## Total DerivativesEdit

The total derivative is important as it preserves some of the key properties of the single variable derivative, most notably the assertion differentiability implies continuity

Let ${\displaystyle f:A\subseteq \mathbb {R} ^{n}\to \mathbb {R} ^{m}}$

We say that ${\displaystyle f}$  is differentiable at ${\displaystyle \mathbf {a} \in A}$  if and only if there exists a linear transformation, ${\displaystyle \mathbf {D} f(\mathbf {a} ):\mathbb {R} ^{n}\to \mathbb {R} ^{m}}$ , called the derivative or total derivative of ${\displaystyle f}$  at ${\displaystyle \mathbf {a} }$ , such that

${\displaystyle \lim _{\|\mathbf {h} \|\to 0}{\frac {\|f(\mathbf {a} +\mathbf {h} )-f(\mathbf {a} )-\mathbf {D} f(\mathbf {a} )(\mathbf {h} )\|}{\|\mathbf {h} \|}}=0}$

One should read ${\displaystyle \mathbf {D} f(\mathbf {a} )(\mathbf {h} )}$  as the linear transformation ${\displaystyle \mathbf {D} f(\mathbf {a} )}$  applied to the vector ${\displaystyle \mathbf {h} }$ . Sometimes it is customary to write this as ${\displaystyle \mathbf {D} f(\mathbf {a} )\cdot (\mathbf {h} )}$ .

### TheoremEdit

Suppose ${\displaystyle A\subseteq \mathbb {R} ^{n}}$  is an open set and ${\displaystyle f:A\to \mathbb {R} ^{m}}$  is differentiable on A. Think of writing ${\displaystyle f}$  in components so ${\displaystyle f(x_{1},\ldots ,x_{n})=(f_{1}(x_{1},\ldots ,x_{n}),\ldots ,f_{m}(x_{1},\ldots ,x_{n}))}$ . Then the partial derivatives ${\displaystyle {\frac {\partial f_{j}}{\partial x_{i}}}}$  exist, and the matrix representing the linear transformation ${\displaystyle \mathbf {D} f(\mathbf {x} )}$  with respect to the standard bases of ${\displaystyle \mathbb {R} ^{n}}$  and ${\displaystyle \mathbb {R} ^{m}}$  is given by the Jacobian Matrix:

${\displaystyle {\begin{bmatrix}{\frac {\partial f_{1}}{\partial x_{1}}}&\cdots &{\frac {\partial f_{1}}{\partial x_{n}}}\\\vdots &\ddots &\vdots \\{\frac {\partial f_{m}}{\partial x_{1}}}&\cdots &{\frac {\partial f_{m}}{\partial x_{n}}}\end{bmatrix}}.}$

evaluated at ${\displaystyle \mathbf {x} =(x_{1},\ldots ,x_{n})}$ .

NOTE: This theorem requires the function to be differentiable to begin with. It is a common mistake to assume that if the partial derivatives exist then this would imply that the function is differentiable because we can construct the Jacobian matrix. This however is completely false. Which brings us to the next theorem:

### TheoremEdit

Suppose ${\displaystyle A\subseteq \mathbb {R} ^{n}}$  is an open set and ${\displaystyle f:A\to \mathbb {R} ^{m}}$ . Think of writing ${\displaystyle f}$  in components so ${\displaystyle f(x_{1},\ldots ,x_{n})=(f_{1}(x_{1},\ldots ,x_{n}),\ldots ,f_{m}(x_{1},\ldots ,x_{n}))}$ . If ${\displaystyle {\frac {\partial f_{j}}{\partial x_{i}}}}$  exists and is continuous on ${\displaystyle A}$  for all ${\displaystyle j\in \{1,\ldots ,m\}}$  and for all ${\displaystyle i\in \{1,\ldots ,n\}}$ , then ${\displaystyle f}$  is differentiable on ${\displaystyle A}$ .

This theorem gives us a nice criteria for a function to be differentiable.