Correlated Gaussian method in Quantum Mechanics

Explicitely correlated Gaussian functions have been extensively used in quantum-mechanical variational calculations in atomic, molecular, and nuclear physics. This book is an attempt to collect the relevant information about this established tool in computational quantum mechanics.



Often we describe bound-state and scattering problems in nuclear and atomic physics through the Shrödinger equation. Unfortunately modern quantum physics offers problems that we can not solve analytically. Luckily the availability of powerful computers is shifting the emphasis from the analytical computation of the solution toward numerical analysis. During the last century numerous methods were developed in order to approximate solutions numerically, e.g. Monte-Carlo simulation, Hypershperical expansion, variational methods with different trial wave functions etc.

In this section we discuss variational method with trial function in the form of correlated Gaussians which is widely used in the modern physics. Mathematically it is based on the Ritz theorem, that states that for an arbitrary function Ψ from the state space the expectation value of the Hamiltonian (<Ψ|H|Ψ>/<Ψ|Ψ>) is larger then the ground state energy. So choosing different trial wave functions and calculating mean values of the Hamiltonian for this functions allows us to get an upper bound for the ground state energy.


To show the idea of the method we consider two particles in 1 dimension interacting through the oscillator potential


It is really simple textbook problem with the ground state solution


where we assumed for simplicity, that the total momentum is equal to zero.

To show how the method works we choose the trial wave function in the form of just one gaussian


where we have only one real positive parameter which has to minimize the energy. The idea of the method is to pick this parameter stochastically using just generator of the real numbers. We find out that independently of the seed after 50 attempts we found value of α that gives the ground state energy with 5 significant digits.

Convergence of the expectation value of the Hamiltonian (different curves corresponds to different starting seeds for the generator of the random numbers).

Of course it's really simple example and we can establish energy with high precision just because we are working in the space that contains ground state wave function of the Hamiltonian.

In order to establish excited states it's not enough to use just one Gaussian so we pick trial wave function in more general form


As above we pick parameters   stochastically and then determine linear parameters ci demanding minimal expectation value of the Hamiltonian. Using N=25 and one random set of the parameters   (we assume that  ) we get first 6 eigenstates with 5 significant digits.

From this simple example we learn that we are able to approximate solution of the Schrödinger equation without any preliminary knowledge about the system using only random search. The main problem is to estimate how good this approximation is.

The methodEdit

Hamiltonian under considerationEdit

We shall consider a non-relativistic quantum mechanical  -particle (N-body) system described by a Hamiltonian


where   and   are the mass and the coordinate of particle number  ; the first term is the kinetic energy operator; the second term is the one-body force, like the external field (often an oscillator,  ), for example Magneto-optical trap; and the third term is the two-body forces, the inter-particle interactions.

This Hamiltonian can describe a system of atoms in a trap or a nucleus.

In general one can also consider spin-dependent, three-body, non-local and other types of interactions which we shall introduce when needed.

Basis expansion of the Schrödinger equationEdit

We are going to solve the Schrödinger equation


where   is the Hamiltonian of a quantum few-body system,   and   are the eigenfunction and the eigenvalue to be found.

We shall expand the wave-function   in terms of a set of basis functions  ,


Inserting the expansion into the Schrödinger equation and multiplying from the left with   gives


Or, in the matrix notation


where   and   are correspondingly the Hamiltonian and the overlap matrices with the matrix elements


The matrix equation   is called the generalized eigenvalue problem. There exist established routines to solve this problem, for example the function "eig" in Octave, or the function "gsl_eigen_gensymm" in the GNU Scientific Library.

If the basis function are orthogonal and normalized the overlap matrix equals the unity matrix,  , and the generalized eigenvalue problem reduces to the ordinary eigenvalue problem,  .

We shall use Gaussians --- which are not orthogonal --- as basis functions, therefore we shall deal with the generalized eigenvalue problem.

Gaussians as basis functionsEdit

We shall use the so-called Correlated Gaussians (or Explicitly Correlated Gaussians) as the basis functions. For a system of  particles with coordinates  , the Correlated Gaussian is defined as


where   denotes the dot-product of the two vectors; and where  , a symmetric positive-defined matrix, and  , the shift-vectors, are (cleverly chosen) parameters of the Gaussian.

In matrix notation,


where   is the column of the coordinates   and   is the column of the shift-vectors  ,




Matrix elements with GaussiansEdit

We represent the few-body wave-function  as a linear combination of   correlated Gaussians,


We shall discuss the choice of matrices   in these Gaussians later and here only calculate the matrix elements.


The overlap,


(where  ,   and  ) can be calculated making an orthogonal transformation,  , where  , to the basis where the matrix   is diagonal,




Kinetic energyEdit

Let us consider a more general form of the kinetic energy operator,


where   is a constant matrix, for example, for the Hamiltonian above  

The matrix element is given as


Integrating by parts with respect to the left derivative gives


Calculating term after term,



where  .



Potential energyEdit

The matrix element of the interaction potential between, say, particles   and  ,


can be written in a more general form,


where   is a size-N vector with all components equal zero except for   and  .

A one-body interaction,  , has the same form,  , where   and all other components equal zero.

Central potentialEdit

For a central potential in a Gaussian form,


the matrix element between shifted Gaussians can be calculated straightforwardly,


where  .

The rank-1 updates of the determinant,  , and the matrix inverse,  , can be efficiently calculated using the following formulae,


and (Sherman-Morrison)


For a general form-factor central potential one way to calculate the matrix element is through the Fourier transform of the potential,  ,




where   is the Fourier transform of the potential  ,   and  .

The last integral can be also written via the potential itself,


where  .



where   and  .

Here are integrals for some popular potentials,

  • Gaussian, ;
  • Coulomb,  ;
  • Oscillator,  .
Tensor potentialEdit

In nuclear physics the tensor potential between two nucleons has the form


where   is the form-factor;  ;   and   are the coordinates of the nucleons;  ,   are the Pauli matrices related to the spins,   and  , of the two nucleons:  ,  .

One often adds the term   to make sure that the potential has no central component (that is, the average of the potential over all directions is zero). Without this extra term the above tensor potential has a central spin-spin component  ).

Again introducing the column   with  ,  ,  , and the vector-columns  ,  , the potential can be written in a convenient general form,


The matrix element of this operator between shifted Gaussians is given as


This can be calculated analytically for a Gaussian form-factor,  ,


where  .

Finally, for a Gaussian tensor potential,


where  .

Spin-orbit potentialEdit

The spin-orbit potential between two nucleons --- with coordinates   and   and spins   and   --- is usually written in the form


where  ;   is the total spin of the two nucleons,


and   is the relative orbital momentum between the two nucleons


where   denotes vector-product of two vectors.

The orbital momentum can be rewritten --- using the column   with  ,  ,   --- in a convenient general form,


For a Gaussian form-factor,  , the corresponding matrix element between shifted Gaussians can be calculated analytically,


where  ,  ,  .

Mathematical formulation for the ground state problemEdit

Let us consider a time-independent physical system whose Hamiltonian H is Hermitian and bounded from below. We want to approximate the discrete eigenvalues of H and its wave functions


where we ordered eigenvalues s.t. 

It means that we would like to find such square integrable functions  , that  , with some  . Unfortunately in practice we don't know exact eigenvalues of the Hamiltonian, so first we have to find approximation to the energy  . The following theorem gives us the receipt. Here we would like to restrict ourselves to the ground state, but using the Min-max theorem one can extend to the whole discrete spectrum of the Hamiltonian


The expectation value of the Hamiltonian for any   from the state space is equal or larger then the ground state energy  .


Apparently the function   can be decomposed in the orthogonal basis  :  . With this decomposition we write the mean value of the Hamiltonian:  , from which follows that  .

This statement is often called Ritz theorem and might be seen as a corollary of the Min-max theorem.

This result allows us compute an upper bound for the ground state energy.

The following theorem according to Weinstein allows us to rewrite our initial demand that   in terms of the variance  


There exist at least one eigenvalue in the interval   .


We write   in the   basis, and get  . There exist integer  , s.t.  . With this we rewrite variance  

This result might be useful if and only if the lower bound can be calculated as close as possible to the ground state energy.

With these theorems we see the way to proceed:

1. Take convenient basis in the state space of the Hamiltonian.

2. Cut the basis size to some finite number.

3. Minimise expectation value of the Hamiltonian in this basis.

4. Enlarge basis and do step 3.

5. Do steps 3,4 as long as needed to insure convergence of the ground state energy.

6. Calculate variance.

7. If variance is larger than some precision value than enlarge basis size and do 3,4,5,6 again, otherwise we are done.

In practice steps 3,4,5 alone can give accurate value of energy. Steps 6,7 are needed for approximation of the wave function. This is due to the following theorem


The expectation value of the Hamiltonian is stationary in the neighbourhood of the discrete eigenvalues.


So in general it is easier to get accurate approximation to the energy than to other observables.


We want to start with the first step: take some convenient basis. We would like to define convenient for our problem

1. Simple transformation from one system of coordinates to another.

2. Possibility to eliminate the centre of mass.

3. Easy computations for the overlap and kinetic energy.


It is of advantage to introduce rescaled coordinates,


where   is a conveniently chosen mass scale. Indeed the kinetic energy   and the harmonic trap   have a more symmetric form in the rescaled coordinates,


The Jacobian of the transformation from   to   is equal


A further suitable linear transformation to a new set of coordinates is possible,


or, in matrix notation,


where   is the transformation matrix.

If the transformation matrix is unitary,  , the diagonal form of the kinetic energy and the harmonic trap is preserved in the new coordinates,


Last transformation is of particular use if new system has the coordinate


which can be seen as a center of mass coordinate. It allows us to work with a wave function in the form


where   is the ground state wave function for the oscillator potential.

Correlated Gaussians General CaseEdit

First we consider trial wave function in the basis of completely general shifted Gaussian, which can be used to describe a system in the external field with anisotropic inter-particle interaction




 , a symmetric positive-defined matrix, and  , a shift vector, are the non-linear parameters of the Gaussian and n=N-1. With this definition we have   non-linear variational parameters. To find those one can use deterministic methods (e.g. Powell's method) or methods based on a stochastic search. We use latter approach though we find   linear variational parameters through a full minimization with respect to a given set of non-linear parameters.

Matrix elementsEdit

The matix elements can be determined analytically either by diagonalizing the matrix   or Cholesky decompose   (since it is positive definite) and change basis to coordinates where the matrix is either diagonal or unit. This way many integrals can be determined by iterative integration.


Correlated Gaussians are generally non-orthogonal and the overlap is therefore non-diagonal,


where we defined  

Kinetic energyEdit

Here we calculate kinetic energy


where we defined   for   with   to be the identity matrix, one can get simpler expression after noticing that


To proceed one has to derive the following identities:


To calculate


which we need to calculate variance we have to calculate the following matrix element


Potential energyEdit

Here we calculate matrix element


In general we can not write analytical expression for this integral, but we can reduce it to D dimensional integrals. For example, consider just one term from the sum


to simplify this integral we have to make transformation from Jacobi set   , where matrices A,A',s,s' are defined to the Jacobi set  , where  . transformation between those sets are provided through the orthogonal matrix U: x=Uy. With this we write




can be found analytically.

If we can write potential as a sum of Gaussians   then the integral   can be found analytically in the same way as we found overlap.



Particles with spinEdit

To consider particles with spin we add spin part to the trial wave function


where for particles with spin = 1/2 function   is just an array of N elements. Each element is an eigenfunction of the spin's projection on the predefined axis. For example   defines the system with all particles have spin in the same direction. Next we define the spin operator   that acts on the particle number   in the following way


and zero otherwise.


Here we discuss spin-orbit potential of the form  , where   and   - relative angular momentum, where   - Levi-Chevita symbol, and  . We have to calculate following matrix element


again we are making transformation from the Jacobi set   to the Jacobi set   using a transformation matrix U: x=Uy.


Correlated Gaussians with Super VectorsEdit

In the previous section we considered completely general setup, which is suitable for any inter-particle potentials and external fields. This approach is far from optimal if for example we are interested in ground state of N bosons with isotropic pairwise interaction, because in this case we know that our ground state must have zero orbital momentum, with this in mind we write the trial wave function in the smaller variational basis:




If we put shift vectors to be zeros  , then the trial wave function treats Cartesian components of vectors   equivalently, which leads to zero angular momentum, otherwise the wave function will contain all possible angular momentum and we need an effective procedure to build an eigenstate for a given angular momentum. Matrix elements for this trial wave functions can be obtained from the general case, but we write it explicitly.

Matrix ElementsEdit



where we defined  

Kinetic energyEdit


where we defined  

Angular momentumEdit

We consider the matrix element of the operator  , choice of   can give the total angular momentum or an angular momentum for appropriate relative coordinate.

First we calculate matrix elements of the form


and now we can calculate matrix element for operator  


We define total angular momentum to be  . If we make transformation to the Jacobi set, than we obtain  , where   is the linear momentum corresponding to the   coordinate. So if we assume that the system as a whole is at rest s.t.   than the following matrix element defines total angular momentum


After simplification (first we rotate to the set of coordinate, where matrix   takes a diagonal form, than we rotate back and rotate to the set where matrix   takes a diagonal form, and again rotate back) we obtain


We take the following integral


where   - is a Kronecker delta.

With this we write total angular momentum



Appendix A: some integrals with 3D GaussiansEdit























Here we need to indroduce a notation to simplify the expressions:








Appendix B: some integrals with 1D GaussiansEdit

First the overlap (denoted N and used in the following expressions)


Polynomial like termsEdit



for any matrix F with the right dimensions including  .



Exponential and sinusoidal termsEdit


(Co)Sine term:

Due to the Euler formulae for (Co)Sine and the linearity we can express these matrix elements as the (Real) Imaginary part of the above matrix element with complex vector  



Gaussian term:


Delta term:

This can be calculated as the limiting case   of the previous matrix element