# Mathematics for Chemistry/Integration

## Free Web Based Material from HEFCE edit

There is a DVD on integration at Math Tutor.

### The basic polynomial edit

This works fine for all powers except -1, for instance the integral of

is just

-1 is clearly going to be a special case because it involves an infinity when and goes to a steep spike as gets small. As you have learned earlier this integral is the natural logarithm of and the infinity exists because the *log* of zero is minus infinity and that of negative numbers is undefined.

### The integration and differentiation of positive and negative powers edit

>>>>>>>>>>>>>>>>>>>>>>Differentiation 1/3 x*x*x x*x 2x 2 0 0 0 0 1/3 x*x*x x*x 2x 2 ? ? ? ? Integration<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<< I(x) H(x) G(x) F(x) ln(x) 1/x -1/(x*x)

Here I, H, G and F are more complicated functions involving .

You will be able to work them out easily when you have done more integration. The thing to notice is that the calculus of negative and positive powers is not symmetrical, essentially caused by the *pole* or *singularity* for at .

## Logarithms edit

Logarithms were invented by Napier, a Scottish Laird, in the 17th-century. He made many inventions but his most enduring came from the necessity of doing the many long divisions and trigonometric calculations used in astronomy. The Royal Navy in later years devoted great time and expense to developing logarithm technology for the purposes of navigation, leading to the commissioning of the first mechanical stored program computer, which was theoretically sound but could not be made by Charles Babbage between 1833 and 1871.

The heart of the system is the observation of the properties of powers:

This means that if we have the *inverse* function of we can change a long division into a subtraction by looking up the exponents in a set of tables. Before the advent of calculators this was the way many calculations were done.

Napier initially used logs to the base for his calculations, but after a year or so he was visited by Briggs who suggested it would be more practical to use base 10. However base is necessary for the purposes of calculus and thermodynamics.

## Integrating 1/x edit

This is true because is our function which reduces or grows at the rate of its own quantity.

This is our definition of a logarithm.

therefore

therefore

## Integrating 1/x like things edit

Just as

so

therefore by the *chain rule*

therefore

Examples of this are:

or

and

As the integral of is so the differential of is so

is just a constant so so

This can also be done by the chain rule

What is interesting here is that the 5 has disappeared completely. The gradient of the log function is unaffected by a multiplier! This is a fundamental property of logs.

## Some observations on infinity edit

Obviously is .

nd are undefined but sometimes a large number over a large number can have defined values. An example is the of 90 degrees, which you will remember has a large opposite over a large hypotenuse but in the limit of an infinitesimally thin triangle they become equal. Therefore the is 1.

## Definite integrals (limits) edit

Remember how we do a definite integral

where is the indefinite integral of .

Here is an example where limits are used to calculate the 3 areas cut out by a quartic equation:

.

We see that is a solution so we can do a polynomial division:

x3 -x2 -2x ------------------- x-1 ) x4 -2x3 -x2 +2x x4 -x3 -------- -x3 -x2 -x3 +x2 ------- -2x2 +2x -2x2 +2x -------- 0

So the equation is which factorises to

.

## Integration by substitution edit

where .

## Simple integration of trigonometric and exponential Functions edit

### Answers edit

*i.e.* the integral of .

## Integration by parts edit

This is done in many textbooks and Wikipedia. Their notation might be different to the one used here, which hopefully is the most clear. You derive the expression by taking the product rule and integrating it. You then make one of the into a product itself to produce the expression.

(all integration with respect to . Remember by

( gets differentiated.)

The important thing is that you have to integrate one expression of the product and differentiate the other. In the basic scheme you integrate the most complicated expression and differentiate the simplest. Each time you do this you generate a new *term* but the function being differentiated goes to zero and the integral is solved. The expression which goes to zero is .

The other common scheme is where the parts formula generates the expression you want on the right of the equals and there are no other integral signs. Then you can rearrange the equation and the integral is solved. This is obviously very useful for trig functions where *ad infinitum*.

also generates itself and is susceptible to the same treatment.

We now have our required integral on both sides of the equation so

## Integration Problems edit

Integrate the following by parts with respect to .

Actually this one can be done quite elegantly by parts, to give a two term expression. Work this one out. Expanding the original integrand by Pascal's Triangle gives:

2 3 4 5 6 7 8 x + 7 x + 21 x + 35 x + 35 x + 21 x + 7 x + x

The two term integral expands to

2 3 4 5 6 7 8 9 1/2 x + 7/3 x + 21/4 x + 7 x + 35/6 x + 3 x + 7/8 x + 1/9 x - 1/72

So one can see it is correct on a term by term basis.

If you integrate you will have to apply *parts* 7 times to get to become 1 thereby generating 8 terms:

7 6 5 4 3 -x cos(x) + 7 x sin(x) + 42 x cos(x) - 210 x sin(x) - 840 x cos(x) + 2 2520 x sin(x) + 5040 x cos(x) - 5040 sin(x) + c (Output from Maple.)

Though it looks nasty there is quite a pattern to this, 7, 7x6,7x6x5 ------7! and sin, cos, -sin, -cos, sin, cos *etc* so it can easily be done by hand.

## Differential equations edit

First order differential equations are covered in many textbooks. They are solved by integration. (First order equations have , second order equations have and .)

The arbitrary constant means another piece of information is needed for complete solution, as with the Newton's Law of Cooling and Half Life examples.

Provided all the s can be got to one side and the s to another the equation is separable.

This is the *general solution*.

Typical examples are:

by definition of logs.

(1)

(2)

(3)

This corresponds to:

The Schrödinger equation is a 2nd order differential equation *e.g.* for the particle in a box

It has taken many decades of work to produce computationally efficient solutions of this equation for polyatomic molecules. Essentially one expands in coefficients of the *atomic orbitals*. Then integrates to make a differential equation a set of numbers, integrals, in a *matrix*. Matrix algebra then finishes the job and finds a solution by solving the resultant simultaneous equations.

## The calculus of trigonometric functions edit

There are many different ways of expressing the same thing in trig functions and very often successful integration depends on recognising a trig identity.

but could also be

(each with an integration constant!).

When applying calculus to these functions it is necessary to spot which is the simplest form for the current manipulation. For integration it often contains a product of a function with its derivative like where integration by substitution is possible.

Where a derivative can be spotted on the numerator and its integral below we will get a function. This is how we integrate .

We can see this function goes to infinity at as it should do.

## Integration by rearrangement edit

Take for example:

Here there is no function producted in with the powers so we cannot use substitution. However there are the two trig identities

and

Using these we have

so we have two simple terms which we can integrate.

## The Maclaurin series edit

We begin by making the assumption that a function can be approximated by an infinite power series in :

By differentiating and setting one gets

Sin, cos and can be expressed by this series approximation

Notice also works for negative .

When differentiated or integrated generates itself!

When differentiated generates .

By using series we can convert a complex function into a polynomial, and can use for small .

In actual fact the kind of approximation used inside computer programs is more like:

These have greater range but are much harder to develop and a bit fiddly on the calculator or to estimate by raw brain power.

We cannot expand this way because is

. However can be expanded.

Work out the series for .

### Factorials edit

The factorials you have seen in series come from repeated differentiation. also has a statistical meaning as it is the number of unique ways you can arrange objects.

is 1 by definition, *i.e.* the number of different ways you can arrange 0 objects is 1.

In statistical thermodynamics you will come across many factorials in expressions such as:

Factorials rapidly get unreasonably large: 6! = 720, 8! = 40320 but 12! = 479001600 so we need to divide them out into reasonable numbers if possible, so for example .

### Stirling's approximation edit

Also in statistical thermodynamics you will find *Stirling's approximation*:

This is proved and discussed in Atkins' *Physical Chemistry*.

How can you use series to estimate . Notice that the series for

converges extremely slowly. is much faster because the

denominator becomes large quickly.

### Trigonometric power series edit

Remember that when you use and

that x must be in radians.....

## Calculus revision edit

### Problems edit

integrate x to the power of x with respect to x

- Differentiate , with respect to . (Hint - use the chain rule.)
- Differentiate . (Chain rule and product rule here.)
- Differentiate . (Hint - split it into a sum of logs first.)
- Integrate . (Hint - use integration by parts and take the expression to be differentiated as 1.)

### Answers edit

- It is just . Bring a out of each term to simplify to .
- .
- - therefore it is 4 times the derivative of .
- You should get by 1 application of
*parts*.