Jump to content

User:Pbehroozi

From Wikipedia, the free encyclopedia

The LCAO Method

[edit]

A simple technique for estimating the ground-state energy of a many-atom system involves using the variational principle with a linear combination of single-atom wavefunctions. While sound in principle, a calculation using this method on the hydrogen molecule ion converges very slowly on the correct binding energy, for the reason that single-atom wavefunctions are not sufficiently dense in the region of overlap.

Introduction

[edit]

The Variational Principle

[edit]

The variational principle is a simple and powerful method for establishing an upper bound on the ground-state energy of a quantum system. Given any wavefunction , we may write as a linear combination of eigenfunctions of the Hamiltonian:

(where, to meet the requirement of normalization, we require ). Then, if we take the inner product of the wavefunction with the Hamiltonian , we find

Letting be the energy associated with each eigenfunction, and letting be the ground state, this becomes

Or, more succinctly,

If we have a large family of wavefunctions, we can calculate for each one of them; with some amount of luck, one member of the family will be close to the actual ground-state wavefunction, and we'll have a good upper estimate for .

The LCAO Technique

[edit]

The LCAO (Linear Combination of Atomic Orbitals) technique uses the variational principle together with a specific family of wavefunctions to find ground-state energies. Specifically, for many-atom systems, each electron will be influenced by a potential

Where refers to the number of protons in each atom and refers to the distance between the electron and each atom. Regardless of other electrons, if we consider locations very close to a particular atomic center (say, for a particular choice of ), we find that the non-relativistic Hamiltonian is

To most significant order, then, this looks exactly like the Hamiltonian for the Hydrogen atom with a charge instead of at the center. As we come closer and closer, the agreement will become better and better as dominates over the rest of the potential terms in the Hamiltonian. If are the wavefunctions for Hydrogen with protons in the center nucleus, then we may well expect that a linear superposition of ground states, namely,

(for some constants ) will match the real ground-state wavefunction to most significant order.

What about higher-order corrections? While the linear combination of ground states is physically motivated, the LCAO technique for higher-order terms is more mathematically motivated. There will be some wave-function which is the ground state of the system; but any wave-function can be decomposed into sums of the eigenfunctions of the Hamiltonian for Hydrogen. Hence, no matter what the real ground-state wave-function looks like near the ith nucleus, we can decompose it into a sum of Hydrogen wave-functions. As this is true near any of the nuclei, we might expect that higher-order corrections to the potential will be reasonably matched by including higher-order wavefunctions of Hydrogen in our linear combination:

But in general, in the limit of including the entire set of eigenfunctions, we will match the ground-state wavefunction. Hence, if the infinite vector space of functions in this family is denoted by , we find that

A Computation Using LCAO

[edit]

Background

[edit]

How well does this work in practice, however? In order to find out, we consider the simplest possible non-Hydrogen system, namely a molecular Hydrogen ion (). For this system, the Hamiltonian is

where and are the distances from the electron to each proton. If we adopt natural units of length corresponding to

(about 0.26 Angstrom) and natural units of energy corresponding to

(about 54.4 eV), the Hamiltonian reduces to the very simple form

If we place both protons on the -axis, separated by a distance , we can write the Hamiltonian in cylindrical coordinates, where it becomes

Written in this way, it becomes explicit that the ground-state energy depends on the separation distance . Hence, without knowing the value of beforehand, we will have to calculate it as well:

To reduce the parameter space to a finite computational problem, we impose a cutoff on , the highest order of the radial wave-functions of hydrogen which we consider. Additionally, we can exploit the symmetries of the Hamiltonian to note that the ground state must satisfy two conditions: (polar symmetry), and (mirror symmetry). Variation from these conditions implies a less-than-minimal value of , which in turn implies a larger value for the energy of the wavefunction.

These two conditions lead to the following restrictions on the coefficients :

The latter two come from the fact that the Legendre polynomials in the wavefunctions of Hydrogen are odd if is odd, and they are even if is even. Hence, while we expect the lowest-order guess for the wavefunction to be of the form

The term involving the odd wavefunction will be antisymmetric:

Finally, one final computational aid may be realized. As the wavefunctions used are eigenfunctions of most of the Hamiltonian---except for the potential from the extra proton, we find that

, and likewise,

, where

is the energy of the th radial state of the Hydrogen atom (in natural units).

With these considerations in effect, we are now ready to perform the variational calculation.

Computational Results

[edit]

Starting with and , we may perform a gradient descent in parameter space to find the lowest energy value of the system. The results of this process are shown here:

Maximal
1 1.32 Å -1.76 eV
2 1.27 Å -1.96 eV
3 1.25 Å -2.07 eV
4 1.24 Å -2.08 eV

In this table, refers to the binding energy of the system, namely, the energy of the electron and both protons relative to the energy of a free proton and a hydrogen atom. The column headed "Maximal " refers to the largest for which was taken to be nonzero.

While these values seem to be converging rapidly, this similitude is misleading. The experimentally-determined value of is -2.65 eV[1], with a separation distance of 1.06Å[2]. This is not an effect of relativistic dynamics (as the electron energy is much less than 511 keV) and is not an effect of fine structure or other corrections (which would modify the energy at the hundredth of a percent level). The full energy of the electron predicted from the last entry in the table is -27.2 eV. Comparatively, the full energy from experiment would be -29.6 eV.

Why is the LCAO method so terrible in this case? A naive guess would be because we have stopped at . However, it is easy to see that adding further orbitals would result in mostly wasted efforts. The th radial eigenfunction of Hydrogen depends on a characteristic radius , where in the natural units we have assumed. However, as the optimal distance between the two protons is about 1.06 Å, or 4.0 in the natural units we have adopted, whereas the natural distance for the wavefunctions is 10, it is easy to see why adding them will have little additional effect. Most of the variation in the wavefunctions will happen at a radius beyond the distance between the two protons, whereas most of the interesting deviation in the ground-state wavefunction happens between the two protons. Hence, beyond the Hydrogen wavefunctions represent a particularly feeble basis with which to describe the ground-state wavefunction of the Hydrogen ion molecule, and so converge very poorly to the correct result.

The moral of this report is that, for calculating energies (or wavefunctions) based on the variational principle, the basis functions of the wavefunction family under consideration must quickly converge to arbitrary functions in the region of overlap between any two independent potentials in the Hamiltonian.

References

[edit]
  1. ^ J. Weiss, Nature, 165, 728 (1950)
  2. ^ The Hydrogen-Ion Molecule

Analytical Imaginary Time Propagation at a Single Point

[edit]

The Imaginary Time Propagation Method (ITP) has long been used as a way to find ground-state wavefunctions.[1] However, if one is only interested in the ground-state energy, a few tricks may be employed to reduce the computational burden. In particular, a method is described whereby for particles with an arbitrary Coulombic potential, the analytical calculations may be done in a -dimensional space instead of a -dimensional space.


Introduction to Imaginary Time Propagation

[edit]

In general, finding the ground-state wavefunction directly from the Hamiltonian for a quantum system is a nightmare, with only a few known exactly-solvable systems. However, the Schrodinger equation yields to a cute mathematical trick, so that it is possible to approximate a ground state wavefunction to arbitrary accuracy. This trick, known as Imaginary Time Propagation (ITP), involves substituting imaginary time instead of real time into the time-dependent Schrodinger equation. This leads to the derivation of an imaginary-time operator which results in exponential decay of all states except for the ground state, which can then be applied to any initial trial wavefunction to compute an approximation to the actual ground state.

As mentioned, we begin with the time-dependent Schrodinger Equation:

Equivalently, in terms of the derivative with respect to imaginary time , we find that

To see why this is useful, we first write as a sum of eigenfunctions of the Hamiltonian , where each eigenfunction has corresponding energy :

For each eigenfunction, the imaginary-time Schrodinger equation then becomes

The solution to this expression is an exponential; if we define , we find that

Where is defined as usual by the Taylor expansion

Evidently, when propagated forward in imaginary time, each eigenfunction decays exponentially, with the rate of decay proportional to its energy. Indeed, all the states other than the ground state will die off exponentially quickly even compared to how quickly the ground state is vanishing:

If we apply the solution for the eigenfunctions to the original wave function, we find

Hence, in the limit of imaginary time propagation, the coefficients for the scattering part of the wavefunction approach zero and those for bound states approach infinity. Note, however, that the relative proportion of the ground state wave function in the overall wave function always increases as a function of imaginary time because :

Thus, starting with any wavefunction , we find that in the limit of large , the wavefunction will become proportional to the ground-state wavefunction:

And, in particular, we may compute the ground-state wavefunction exactly, as long as the overlap between our initial wavefunction and the ground-state wavefunction is nonzero:

The Ground-State Energy

[edit]

ITP at a Single Point

[edit]

Full ITP normally involves numerically propagating a trial wavefunction for many time steps over a large volume in parameter space. If we are only interested in finding the ground-state energy of a system, this seems particularly wasteful. For an eigenfunction of the Hamiltonian, it is true at any point in parameter space that . This suggests a simple method for ITP at a single point. Suppose that in natural units our Hamiltonian for particles is of the form

Where, as usual, is the sum over all (rectilinear) space dimensions of second derivatives:

and

As rectilinear partial derivatives commute with themselves as well as partials with respect to imaginary time (), we may then write down the equation for the imaginary time evolution of the derivatives of :

Now, if we start with the initial value of at a point and some large number of its derivatives, we can evolve both and its derivatives according to the imaginary-time Schrodinger equation. After some appropriate amount of time, we can then extract the ground-state energy just by applying to at the particular point we have chosen.

However, this is not especially better than a brute-force solution. Even though derivatives commute, if there are dimensions, there are distinct ways to take derivatives. While we would not have to worry about spatial boundary conditions, we would have to worry about "boundary" conditions on the derivatives which we choose not to propagate. Hence, we have simply mapped the complexity of the problem in a virtually one-to-one manner into a different (but no more useful) form.

A Better Approach

[edit]

For a better approach, we recall the expression for the ITP of , decomposed into eigenfunctions:

Evidently, if we can calculate the effect of the operator on , we can immediately recover the ground-state energy:

Then, we simply have to calculate a number of terms in the expansion of . How many? The answer naturally depends on the initial wave function which we choose and the desired accuracy which we want. Suppose, for instance, we pick a wavefunction which is some (unknown) superposition of bound states and we desire an accuracy of one part in . We expect that (unless we get extremely unlucky with our choice) we will find

This would suggest that we would like , or, in other words, . In that case, how fast does the series for converge? If we note that , then

Thus, for our desired accuracy, we want

for .

We expect this to be maximum for , if the states are ordered in terms of increasing energy. For Hydrogen, for instance, where , this implies that about 56 terms in the expansion are necessary. For Helium, where , this would imply the calculation of 320 terms. If one had an estimate of , one could naturally take away from the entire energy scale, dramatically reducing the number of terms necessary (38 terms for Hydrogen, 173 terms for Helium).

This, at least, sets a sensible upper bound on the size of the calculation. For any one-dimensional calculation, this method is clearly a fine choice---as each derivative in the Hamiltonian can only be taken one way, calculating each successive term in the expansion only requires that one derivative be taken. However, one-dimensional calculations are hardly impressive. In fact, up to the present time, we have gained no advantages by evaluating at a single point. Since a new Laplacian is applied for each term, we end up taking derivatives to calculate terms, resulting in a computational burden of about distinct derivatives at the final step.

While one could pick a simple initial wavefunction with simple derivatives (such as ), this does not help matters. Observe, for example,

The last term is the trouble, for it means that no matter how simple our wavefunction is, we still end up calculating derivatives of and, by extension, , , and so on.

Coulombic Potentials

[edit]

Nonetheless, with certain potentials, the derivatives of can still end up taking extremely simple forms. One case in point is Coulombic potentials, where

Most obviously, (except at certain points) which reduces the number of distinct derivatives one must calculate. However, there's a more subtle consideration which is much more important. Consider a two-particle potential, with one particle at the origin:

If we evaluate at the point , then any elements of our answer which involve positive powers of or will evaluate to zero, and we can skip calculating them entirely. Indeed, as it turns out, we can skip calculating derivatives with respect to entirely until the very end of our calculation, because they all take a very simple form. In particular, we can make a simple table of even derivatives, noting that all odd derivatives will evaluate to zero at :

Where is a simple function purely depending on : , and where is the semifactorial: for even , and for odd integers.

Evidently, we can apply this to derivatives with respect to , where we find

Or, in general, if we need to make up total derivatives this way, we find

where

If, for whatever reason, we are examining the two-particle system in higher-dimensional spacetime, then further coefficients in the style of are no harder to calculate:

So that calculations of this type are linear in terms of the number of dimensions involved.

Hence, we may treat this as a purely one-dimensional problem, taking only derivatives with respect to , and then only at the very end taking derivatives with respect to and as necessary to round out the Laplacians. That is to say, the effect of adding extra dimensions is a purely combinatorical factor multiplying the final result. Now, of course, this should not be impressive yet---after all, we started with a 1D-potential, namely .

However, with more particles, this trick---namely, that the derivatives all take the same general form---becomes astoundingly useful. We cannot expect to place all the particles on top of each other, obviously. Nonetheless, in problems where the particle locations are not constrained, we choose to evaluate at a location in parameter space where all the particles are lined up in a straight line along the z-axis. That is to say, for each particle, so that and for any pair of particles. Then, we find that the only nontrivial derivatives in the expansion of ---i.e., terms which are not purely powers of the distances between particles---result from derivatives with respect to the particle locations on the z-axis. At the end, we may apply the effects of the extra dimensions all at once---as they cannot result in new terms, and as they only result in multiplication by a simply-calculable number, their computational cost becomes linear in the number of nontrivial derivatives already calculated.

Consequently, we have reduced a -dimensional problem to a -dimensional problem. Hydrogen, as before, remains a 1D-problem; Helium becomes a 2D problem; Lithium becomes a 3D problem, and so on and so forth.

Future Work

[edit]

In the previous section, we only considered fixing the locations of particles in two dimensions each. We consider it likely that by (for example) fixing particles to be at a constant spacing along the axis, it may be possible write all the derivatives along the axis in terms of each other---resulting in the derivative of only one dimension being unique. This would be truly satisfying---as energy is a single-dimensional quantity---we would then be at a minimum of work required to calculate it.

References

[edit]
  1. ^ W. Magnus, Comm. Pure Appl. Math. 7, 649-673 (1954)