try ai
Popular Science
Edit
Share
Feedback
  • Linear Variational Method

Linear Variational Method

SciencePediaSciencePedia
Key Takeaways
  • The variational principle guarantees that the energy calculated from any trial wavefunction is always an upper bound to the true ground-state energy of a system.
  • The linear variational method systematically finds the best approximate solution by constructing a trial wavefunction as a mixture (a linear combination) of simpler, known basis functions.
  • Solving the secular equation yields not only an approximation for the ground state but also a whole spectrum of upper-bound energies for the system's excited states.
  • This method is a foundational tool in quantum chemistry (LCAO, CI) and solid-state physics, providing the basis for understanding chemical bonds, electron correlation, and material properties like conductivity.

Introduction

In the realm of quantum mechanics, the Schrödinger equation holds the key to understanding the behavior of atoms and molecules. However, its exact solution is achievable for only the simplest systems, leaving the vast majority of chemical and physical reality mathematically inaccessible. This gap necessitates the use of approximation methods, and among the most powerful and elegant is the linear variational method. It provides a systematic and reliable way to find approximate solutions for complex quantum systems, turning an unsolvable differential equation into a manageable algebraic problem.

This article delves into this cornerstone of computational science. The first part, "Principles and Mechanisms," will unpack the foundational variational principle and explain how a trial wavefunction can be constructed from a linear combination of simpler functions, leading to the pivotal secular equation. The second part, "Applications and Interdisciplinary Connections," will showcase the method's incredible versatility, demonstrating how this single concept explains everything from the nature of the chemical bond to the electronic properties of solid materials.

Principles and Mechanisms

The grand aim of quantum mechanics, in many ways, is to answer a single question for any given physical system: What are its allowed energy states? The answer lies locked within the system's Schrödinger equation. But, as nature would have it, this equation is notoriously difficult to solve exactly for anything more complex than a hydrogen atom. So, what's a physicist or chemist to do? Do we give up? Not at all. We do what clever people have always done when faced with an impossible problem: we find a brilliant way to cheat. This "cheat" is one of the most elegant and powerful ideas in all of quantum physics: the ​​variational principle​​.

The Search for the Lowest Rung: The Variational Principle

Imagine you are trying to find the absolute lowest point in a vast, fog-shrouded valley. You can't see the entire landscape, so you can't just point to the bottom. However, you have an altimeter that tells you your current elevation. The variational principle is the simple, yet profound, statement that no matter where you stand in the valley, your current elevation is at or above the elevation of the lowest point. You can never find yourself standing at an altitude below the valley floor.

In quantum mechanics, the "valley" is the space of all possible wavefunctions, and the "elevation" is the energy corresponding to a given wavefunction. The variational principle states that for any plausible, well-behaved "trial" wavefunction, ψtrial\psi_{trial}ψtrial​, that we might guess for a system, the expectation value of its energy, E\mathcal{E}E, will always be greater than or equal to the true ground-state energy, E0E_0E0​.

E=⟨ψtrial∣H^∣ψtrial⟩⟨ψtrial∣ψtrial⟩≥E0\mathcal{E} = \frac{\langle \psi_{trial} | \hat{H} | \psi_{trial} \rangle}{\langle \psi_{trial} | \psi_{trial} \rangle} \ge E_0E=⟨ψtrial​∣ψtrial​⟩⟨ψtrial​∣H^∣ψtrial​⟩​≥E0​

The only way to have equality, E=E0\mathcal{E} = E_0E=E0​, is if your guess happens to be the exact ground-state wavefunction. Otherwise, your energy is guaranteed to be an upper bound. Our goal, then, is to intelligently "walk downhill"—to vary our trial wavefunction in a way that minimizes this energy, getting us as close as possible to the true ground-state value.

This isn't just a useful guideline; it's a rigid mathematical theorem. It acts as a stern referee for our calculations. If a student performs a complex computational calculation for the Helium atom and obtains a ground-state energy of −2.9050-2.9050−2.9050 Hartrees, while the experimentally confirmed true value is −2.9037-2.9037−2.9037 Hartrees, we know something is wrong. The calculation gave an energy below the true ground state, which the variational principle forbids. This signals not a breakthrough discovery, but a bug in the code or an error in the method's application. This principle provides a powerful, built-in check on reality.

A Recipe for a Wavefunction: The Art of Linear Combinations

So, how do we make an intelligent guess? We could randomly try functions, but that's like wandering blindfolded in our foggy valley. The ​​linear variational method​​ gives us a systematic strategy. The idea is wonderfully intuitive: we construct our trial wavefunction by mixing together a handful of simpler, well-understood functions, called ​​basis functions​​.

Think of it like an artist trying to match a specific, complex color. They start with a palette of primary colors and mix them in different proportions. In our case, the "colors" are our basis functions (ϕ1,ϕ2,…,ϕn\phi_1, \phi_2, \dots, \phi_nϕ1​,ϕ2​,…,ϕn​), and our "recipe" is a linear combination:

Ψtrial=c1ϕ1+c2ϕ2+⋯+cnϕn\Psi_{trial} = c_1 \phi_1 + c_2 \phi_2 + \dots + c_n \phi_nΨtrial​=c1​ϕ1​+c2​ϕ2​+⋯+cn​ϕn​

The basis functions, ϕi\phi_iϕi​, are chosen for good reasons—perhaps they are atomic orbitals on the atoms in a molecule, or simple polynomials that obey the physical constraints of the problem. They form our fixed "palette." The magic lies in the coefficients, cic_ici​. These are the variational parameters—the "dials" we can turn to adjust the mix, seeking the combination that yields the lowest possible energy.

The Mathematical Engine: Turning Dials with Determinants

How do we find the best settings for our dials? We don't have to guess. The principles of calculus allow us to find the minimum energy by requiring that the derivative of the energy with respect to each coefficient be zero. This procedure beautifully transforms the problem. Instead of wrestling with the fearsome Schrödinger differential equation, we find ourselves needing to solve a set of simultaneous linear equations, known as the ​​secular equations​​.

For these equations to have a meaningful, non-zero solution for our coefficients, a special condition must be satisfied: the determinant of a specific matrix must be zero. This gives us the famous ​​secular equation​​:

det⁡(H−ES)=0\det(\mathbf{H} - E\mathbf{S}) = 0det(H−ES)=0

This equation may look imposing, but it's the heart of the machine. Let's break it down:

  • EEE is the energy we are trying to find.
  • H\mathbf{H}H is the ​​Hamiltonian matrix​​. Its elements, Hij=⟨ϕi∣H^∣ϕj⟩H_{ij} = \langle \phi_i | \hat{H} | \phi_j \rangleHij​=⟨ϕi​∣H^∣ϕj​⟩, represent the energy of a basis function (HiiH_{ii}Hii​) and the energy of interaction between two different basis functions (HijH_{ij}Hij​). It encodes the physics of the problem.
  • S\mathbf{S}S is the ​​overlap matrix​​. Its elements, Sij=⟨ϕi∣ϕj⟩S_{ij} = \langle \phi_i | \phi_j \rangleSij​=⟨ϕi​∣ϕj​⟩, measure the extent to which two basis functions occupy the same space—how much they "overlap". If our basis functions are orthonormal, S\mathbf{S}S is just the identity matrix.

For a simple two-function basis, this determinant expands into a straightforward quadratic equation in EEE. For a larger basis, it becomes a higher-order polynomial. The roots of this polynomial are our calculated energies. The problem of solving the Schrödinger equation has been reduced to the much more familiar task of finding the roots of a polynomial! For a simple model of a molecule, we can write down the H and S matrices based on chemical intuition and solve this determinant to find the allowed molecular orbital energies.

A Ladder of Approximations: Interpreting the Results

When we solve the secular equation for a basis of nnn functions, we don't get just one energy—we get nnn distinct energy values. What do they all mean?

The lowest energy root is our best approximation for the ground-state energy, E0E_0E0​. By the variational principle, we know it is an upper bound to the true value.

But what about the other n−1n-1n−1 roots? Herein lies a fantastic bonus of the method. The second-lowest root turns out to be an upper bound for the true first excited state energy, E1E_1E1​. The third-lowest is an upper bound for E2E_2E2​, and so on. This remarkable property, formally known as the Hylleraas-Undheim-MacDonald theorem, means that our single calculation gives us not just a picture of the ground floor, but an approximate blueprint of the entire energy ladder of the system. We get a whole spectrum of approximate energy levels from a single, unified procedure.

The Path to Perfection: On the Choice of Basis

The accuracy of our results depends entirely on the flexibility of our trial wavefunction, which in turn depends on our choice of basis functions. A poor basis set is like trying to paint a photorealistic portrait using only red, yellow, and blue. You can get a recognizable image, but you'll miss all the subtlety.

So, how do we improve our approximation? The answer is simple and systematic: we expand our basis set. We add more functions to the mix. A crucial feature of the linear variational method is that adding another function to your basis set can never make the calculated ground-state energy worse (i.e., higher). It will either stay the same or, more likely, get better—it will decrease, moving closer to the true value.

Imagine again our foggy valley. A two-function basis is like being constrained to walk along a single straight line. The lowest point you find on that line is your best guess. Now, if you add a third basis function, it's like you're now free to explore an entire two-dimensional plane. The lowest point in this plane is guaranteed to be at least as low as, and almost certainly lower than, the lowest point on your original line.

This provides a clear and powerful way to assess and improve our calculations. If one research group performs a calculation with a simple basis set and gets a ground state energy of, say, −1.333-1.333−1.333 Hartrees, and another group uses a more sophisticated basis and gets −1.435-1.435−1.435 Hartrees, we can say with confidence that the second result is the better approximation. The lower energy wins.

This "path to perfection" tells us that if we could, in principle, use a mathematically "complete" basis set—one that could describe any possible shape—the linear variational method would yield the exact energies. While we can never use an infinite basis in practice, this gives us profound confidence. The method is not just a trick; it's a systematic, improvable pathway that leads us ever closer to the true quantum reality of molecules and materials, from simple chains of atoms to the complex electronic structures that govern our world.

Applications and Interdisciplinary Connections

We have spent some time learning the formal machinery of the linear variational method. It is a beautiful piece of mathematics, elegant and self-contained. But a question should always be nagging at the back of our minds: What is it for? A tool is only as good as the things it can build. We are now in a position to see that this simple idea—of guessing an answer by mixing together simpler, known pieces—is one of the most powerful and versatile tools in the modern physicist's and chemist's toolkit. It is a key that unlocks the secrets of molecules, crystals, and magnets.

The game is always the same: we face a complex system whose exact solution is beyond our grasp. So, we make an educated guess. We propose that the true, complicated answer can be approximated by a linear combination of simpler, more manageable functions that we already understand. The linear variational method is then the impartial referee that tells us the very best way to mix our simple ingredients to get as close as possible to the real thing, as measured by the energy. Let us now embark on a journey to see this principle in action, and we will discover it is the common thread weaving through vast and seemingly disconnected areas of science.

The Heart of Chemistry: Building Molecules

What is a molecule? It is, of course, a collection of atoms. But it is much more than just atoms sitting near each other. A strange and wonderful transformation occurs when atoms bind together; their individual identities blur as their electrons are shared, forming a new, unified quantum entity. The linear variational method provides our most fundamental description of this process, known as the ​​Linear Combination of Atomic Orbitals (LCAO)​​ method.

Let's take the simplest possible molecule: the hydrogen molecular ion, H2+H_2^+H2+​, which is just two protons sharing a single electron. How can we describe the state of this electron? A brilliant and simple guess is to say that the molecular orbital must look something like a mixture of the atomic orbitals of the two hydrogen atoms it came from. Our basis is thus beautifully simple: a 1s atomic orbital centered on proton A, and another 1s orbital centered on proton B.

When we turn the crank of the linear variational method on this two-state basis, something remarkable happens. The secular equations yield not one, but two possible solutions. One solution has an energy lower than the original atomic orbital. In this state, the wavefunction is an additive combination of the two atomic orbitals, causing the electron density to pile up between the two protons. This buildup of negative charge acts as a sort of electrostatic glue, pulling the two positive protons together. We call this a ​​bonding orbital​​. The other solution has an energy higher than the original atomic orbital. Here, the wavefunctions subtract, creating a node—a region of zero electron density—right between the protons. With no glue to hold them, the protons repel each other. This is an ​​antibonding orbital​​.

There it is! The origin of the chemical bond, laid bare. It is not some mystical force, but simply the result of the universe, through the rules of quantum mechanics, seeking the lowest possible energy configuration available by mixing atomic states. This same principle of mixing atomic orbitals is the foundation upon which the entirety of modern quantum chemistry is built, allowing us to understand the structure of molecules as complex as proteins and DNA.

Of course, in real calculations, chemists must be careful about their choice of "simple pieces." Real atomic orbitals (called Slater-Type Orbitals, or STOs) have a sharp "cusp" at the nucleus and decay exponentially at long distances. These features are mathematically difficult to work with when you have many atoms. So, a clever trick is employed: instead of using the physically "correct" but computationally nightmarish STOs, chemists use a basis of Gaussian-Type Orbitals (GTOs). A single GTO is a poor imitation of an atomic orbital—it has no cusp and decays too quickly. But its mathematical properties are wonderfully simple, especially the fact that the product of two Gaussians is another Gaussian. This allows for the efficient calculation of the millions of integrals needed for a large molecule. By combining many GTOs, one can build up a very accurate approximation of the "true" atomic orbital shape. This is a beautiful example of a pragmatic trade-off between physical fidelity and computational feasibility, a constant theme in computational science.

Beyond the Simplest Picture: The Intricate Dance of Electrons

The LCAO model is a monumental success, but it is still an approximation. It generally treats each electron as moving in an average field created by all the other particles. But electrons are not so polite; they are charged particles that actively repel and avoid one another. This complex, correlated motion is one of the most difficult problems in quantum chemistry. How can we improve our description? Once again, we call upon the linear variational method.

The new strategy is called ​​Configuration Interaction (CI)​​. We start with our simple LCAO or Hartree-Fock description of the ground state, which we can think of as one specific arrangement—or "configuration"—of electrons in the available orbitals. We then acknowledge that this cannot be the whole story. The true ground state might contain a small mixture of other, higher-energy configurations, where one or two electrons have been "excited" into what were previously empty orbitals.

Consider the Helium atom. The simplest picture has both electrons in the 1s orbital, a configuration we denote (1s)2(1s)^2(1s)2. To improve this, we can build a variational trial function that is mostly the (1s)2(1s)^2(1s)2 configuration, but with a little bit of, say, the doubly-excited (2p)2(2p)^2(2p)2 configuration mixed in. Our basis now consists of two (or more) many-electron configuration state functions. The linear variational method then tells us the optimal mixture that minimizes the energy. By mixing in these excited configurations, we are explicitly allowing the electrons to arrange themselves in more complex ways to better avoid each other, thus capturing the effects of electron correlation.

A profound advantage of this approach is its mathematical rigor. Because the linear variational method is a direct application of the Rayleigh-Ritz variational principle, any energy we calculate is guaranteed to be an upper bound to the true ground state energy. The more configurations we include in our basis, the better our approximation becomes, and the calculated energy gets closer and closer to the true value, always approaching it from above. This provides a systematic way to improve our calculations and a guarantee that we are never "too good to be true." This stands in contrast to other powerful methods in quantum chemistry, like coupled-cluster theory, which are often more efficient but do not provide this strict variational guarantee.

From Molecules to Materials: The World of Solids and Spins

The power of our method is not confined to the scale of individual molecules. The very same ideas that explain a single chemical bond can also explain the collective properties of a macroscopic crystal containing 102310^{23}1023 atoms.

One of the most fundamental properties of a solid material is whether it conducts electricity or not. This is determined by the allowed energy levels for electrons within the crystal. In a crystalline solid, an electron no longer belongs to a single atom but moves through a periodic potential created by the entire lattice of atomic nuclei. We can model this using the linear variational method in an approach called the ​​nearly-free electron model​​.

Imagine an electron moving through a one-dimensional crystal. We start by guessing its wavefunction is a simple plane wave, like that of a completely free particle. Now, at certain special wavelengths—those that are an integer multiple of the lattice spacing—something interesting happens. A plane wave moving to the right, eikxe^{ikx}eikx, and one moving to the left, e−ikxe^{-ikx}e−ikx, have the exact same energy. They are degenerate. But the moment we turn on even a vanishingly small periodic potential from the crystal lattice, these two states are no longer independent. The potential mixes them.

By setting up a simple 2×22 \times 22×2 variational problem with these two plane waves as our basis, we find that the original degeneracy is lifted. The two states mix to form two new states, one with a slightly lower energy and one with a slightly higher energy. The energy difference between them is a "band gap"—a forbidden range of energies that no electron in the crystal can possess. This simple result is the origin of the distinction between metals (which have no gap at the Fermi level), insulators (which have a large gap), and semiconductors (which have a small gap). The vast technological world of electronics rests on this quantum mechanical splitting of degenerate states.

The method's reach extends even further, into the abstract world of quantum magnetism. Consider a magnetic material described by the Heisenberg model, where atomic spins interact with their neighbors. A fascinating scenario arises in systems with "frustration," such as three spins on a triangle that each want to be anti-aligned with their neighbors. It's an impossible task! If spin 1 is "up" and spin 2 is "down," what should spin 3 do? It cannot be anti-aligned with both.

To find the true ground state of such a system, we can use the linear variational method. Our basis is no longer a set of spatial functions, but a set of all possible spin configurations for the system (e.g., ∣↑↓↓⟩|\uparrow\downarrow\downarrow\rangle∣↑↓↓⟩, ∣↓↑↓⟩|\downarrow\uparrow\downarrow\rangle∣↓↑↓⟩, and ∣↓↓↑⟩|\downarrow\downarrow\uparrow\rangle∣↓↓↑⟩). The Hamiltonian matrix elements describe the energy cost of the spin-spin interactions. Diagonalizing this matrix gives the true energy eigenstates, which are not simple spin arrangements but complex quantum superpositions of all the basis states. This allows us to probe the exotic quantum ground states that emerge from magnetic frustration.

Responding to the World: Calculating Properties

So far, we have used the variational method to find the stationary states and energies of isolated systems. But its utility does not end there. We can also use it to ask how a system responds when we poke it with an external probe, such as an electric field.

Let's return to our H2+H_2^+H2+​ molecule and place it in a uniform electric field. The electron cloud, which was symmetric, will now be distorted, shifting slightly in response to the field. The measure of how much it shifts for a given field strength is the molecule's ​​polarizability​​. To calculate this, we add a new term to our Hamiltonian representing the interaction with the electric field. The matrix elements in our LCAO basis now depend on the field strength, FFF.

We solve the secular equations again, this time finding energy levels that are themselves functions of FFF. The ground state energy in the presence of the field will have a term proportional to F2F^2F2. The coefficient of this term is directly related to the polarizability. By analyzing how the variational energy changes as we turn on the field, we can extract this important physical property.

A Unifying Thread

From the glue holding molecules together, to the intricate dance of electrons avoiding each other, to the origin of conductivity in metals, to the strange behavior of frustrated magnets, and even to how a molecule lights up in an electric field—we have seen the same fundamental idea at work. The linear variational method is a testament to the profound unity of physics and chemistry. It teaches us that we can understand the most complex systems by intelligently combining simple pieces. It is a simple, beautiful, and astonishingly powerful principle for exploring the quantum world.