try ai
Popular Science
Edit
Share
Feedback
  • Statistical Information

Statistical Information

SciencePediaSciencePedia
Key Takeaways
  • The Data Processing Inequality is a fundamental theorem stating that processing data can never create new information about its original source; it can only preserve or lose it.
  • A sufficient statistic is a powerful form of data compression that summarizes a vast dataset into a few key numbers while retaining all the information about a specific parameter of interest.
  • When perfect models are intractable, methods like Approximate Bayesian Computation (ABC) use plausible but insufficient summary statistics to make robust inferences.
  • The concept of statistical information unifies disparate scientific fields, revealing deep connections such as the mathematical equivalence between Fisher information and kinetic energy in quantum chemistry.

Introduction

In the vast ocean of data that the world presents, from the output of a particle accelerator to the sequence of a genome, lies the challenge of distinguishing meaningful clues from random noise. The science of this distinction revolves around the concept of ​​statistical information​​, a profound idea that quantifies what our data can truly tell us about the hidden reality we aim to understand. This is not merely about counting bytes but about extracting the essential signal that points toward scientific truth. This article addresses the fundamental problem of how to identify, preserve, and utilize this information in the face of overwhelming complexity and inevitable loss.

Across the following chapters, you will embark on a journey into the core of this concept. First, we will explore the "Principles and Mechanisms" that govern statistical information, from the foundational limits set by the Data Processing Inequality to the elegant efficiency of sufficient statistics. We will examine the unavoidable costs of information loss and the subtle but crucial differences between various types of information. Following this theoretical grounding, we will witness these ideas in action in "Applications and Interdisciplinary Connections," discovering how statistical information serves as a common language that connects genomics, evolutionary history, and even the fundamental laws of physics, enabling us to decode the patterns of the natural world.

Principles and Mechanisms

Imagine you are a detective at the scene of a crime. The room is filled with countless details: a knocked-over vase, a faint scent of perfume, a half-empty cup of tea, footprints on the rug. To a novice, it's a bewildering chaos of "data." But the master detective knows that only a few of these details are true clues—pieces of information that point toward the truth of what happened. Most of the rest is just noise, distractions from the central question.

Science is much like this. The world bombards us with data, whether from a telescope, a gene sequencer, or a particle accelerator. Our job is to find the clues. In science, we call this the search for ​​statistical information​​. It's a concept far more profound than just counting bytes on a hard drive. It's about quantifying what our data can tell us about the underlying, hidden reality we are trying to understand. This chapter is a journey into the heart of this concept, exploring how we find the essential signal within the overwhelming noise, and what it costs us when we can't.

The Invisible Thread: Data and the Reality It Describes

Let's begin with a simple, but powerful idea. The data we collect is not the reality itself; it is a message, a shadow, a footprint left by reality. Consider the state of a nation's economy—a vastly complex entity we can call XXX. No one can see XXX in its entirety. Instead, the government collects data on employment, inflation, and trade, producing a set of official statistics, YYY. These statistics are a function of the true economy, but they are simplified, aggregated, and perhaps even contain errors. A private analyst then takes these public statistics YYY and creates a forecast, ZZZ.

The analyst's forecast ZZZ is a processed version of the government's data YYY, which itself is a processed version of the true economic state XXX. We have a chain: X→Y→ZX \to Y \to ZX→Y→Z. It seems intuitively obvious, and it is a fundamental theorem of information theory, that the analyst's forecast can never contain more information about the true state of the economy than the government statistics it was based on. And those statistics can't contain more information than the economy itself. This is the ​​Data Processing Inequality​​. Mathematically, it states that the mutual information between the source and the final output is less than or equal to the information between the source and the intermediate step: I(X;Z)≤I(X;Y)I(X; Z) \le I(X; Y)I(X;Z)≤I(X;Y).

You can't get more out than you put in. Processing data—summarizing, filtering, modeling—cannot create information about the original source. At best, it preserves it; more often, it loses it. This single principle is the guiding light and the fundamental constraint in all of data science. Our goal is to process data in such a way that we lose as little information as possible about the question we are asking.

Extracting Gold: From Raw Signals to Meaningful Numbers

Before we can even think about losing information, we must first figure out how to extract it. Raw instrumental output is rarely the information itself. Information is encoded in the features of the data, and we must know the code.

Imagine you're a structural biologist using Nuclear Magnetic Resonance (NMR) to study a protein. The machine gives you a complex spectrum full of peaks. What do these peaks mean? The position of a peak (its chemical shift) tells you about the local chemical environment of a proton. The splitting of a peak (multiplicity) tells you about its neighbors. The width of the peak tells you about its motion. But if you want to know how many protons are contributing to that signal, you must measure the ​​integrated area under the peak​​. This area is directly proportional to the number of protons giving rise to the signal. Out of all the features in that complex spectrum, the area is the specific "statistic" that encodes the count.

Sometimes, this encoding is extraordinarily clever. In modern proteomics, scientists compare the amounts of thousands of proteins between a healthy cell and a diseased one. A technique using Tandem Mass Tags (TMT) attaches a special chemical label to every peptide. These labels are ​​isobaric​​, meaning they all have the same total mass. So, when the peptides are first weighed in the mass spectrometer (in the MS1 scan), a peptide from the healthy sample and the same peptide from the diseased sample are indistinguishable; they appear as a single peak. It seems like the quantitative information is lost!

But here is the trick: upon fragmentation (in the MS2 scan), the tags break apart, releasing small "reporter ions." The masses of these reporter ions are different for each sample. The ratio of the intensities of these reporter ions reveals the relative abundance of the peptide in the original samples. This is a brilliant experimental design. The information is intentionally hidden in one stage of the experiment (MS1) only to be cleanly revealed in the next (MS2). It's like writing a secret message in invisible ink.

Of course, just getting a number isn't enough; we need to trust it. In X-ray crystallography, scientists bombard a protein crystal with X-rays and measure the intensities of thousands of diffracted spots. The quality of the final 3D structure depends critically on the quality of this data. Two key metrics are ​​completeness​​ and ​​redundancy​​. Completeness measures what fraction of all theoretically possible reflections were actually measured. A dataset with 98% completeness is far better than one with 85% because it provides a more comprehensive view of the structure. Redundancy (or multiplicity) measures how many times, on average, each unique reflection was measured. A high redundancy (e.g., 4.7) is better than a low one (e.g., 4.1) because it allows the scientist to average out random measurement errors, leading to more reliable intensity values. High-quality information is not just comprehensive, it's also reliable.

The Alchemist's Dream: Sufficient Statistics

This brings us to one of the most beautiful and powerful ideas in all of statistics: the ​​sufficient statistic​​. A sufficient statistic is a function of the data that contains all the information about the parameter of interest that was present in the original, full dataset. It is the ultimate act of data compression, a magical distillation of a vast and messy dataset into a few numbers with zero loss of information for your specific question.

Let's make this concrete. Imagine you are an ecologist studying a closed population of rare birds in a forest, and you want to estimate the total population size, NNN. You conduct a ​​mark-recapture​​ study over KKK weeks. Each week you capture some birds, tag any untagged ones, record their IDs, and release them. At the end, your field notebook contains a massive dataset of individual capture histories: bird #34 was caught in week 1 and 4; bird #7 was caught only in week 3; and so on.

To estimate the total population size NNN and the probability of capture ppp, do you need all these intricate details? The surprising answer is no. The theory of sufficiency tells us that all of the information about NNN and ppp is contained in a much simpler set of numbers: the total number of unique birds ever seen, n⋅n_{\cdot}n⋅​, and the counts of how many birds were seen exactly once (f1f_1f1​), exactly twice (f2f_2f2​), and so on, up to fKf_KfK​. Whether bird #34 was caught in week 1 and 4, or in week 2 and 3, makes no difference to the estimation of NNN. By reducing the complex histories to these simple counts, we have achieved enormous data compression with no loss of information. This is the alchemist's dream: turning the lead of raw data into the gold of a sufficient statistic.

This principle is what makes modern big data science tractable. A Genome-Wide Association Study (GWAS) might examine the DNA of a million people to find genetic variants associated with a disease. The raw dataset is petabytes in size. Yet, for many purposes, the entire dataset can be summarized. For each genetic variant, we can compute its estimated effect size (β^\hat{\beta}β^​), the standard error of that estimate (SE^\widehat{\mathrm{SE}}SE), and its frequency in the population (ppp). This handful of numbers per variant forms a set of ​​summary statistics​​. Amazingly, this small summary file is sufficient for a vast array of downstream analyses, like combining results from multiple studies (meta-analysis) or fine-mapping the causal gene in a region. Without the principle of sufficiency, collaborative big-data genomics would be nearly impossible.

The existence of a sufficient statistic often stems from the underlying mathematical structure of the physical model. In a study of chemical reactions, if you want to distinguish a "stripping" mechanism from a "rebound" or "harpoon" mechanism, and you measure the scattering angle, product energy, and charge transfer for thousands of individual reactions, you don't need to keep the full list of thousands of measurements. If your model belongs to a common class known as an ​​exponential family​​, the minimal sufficient statistic is simply the triplet of sums: the sum of the cosines of the scattering angles, the sum of the energies, and the sum of the charge transfer indicators. All the information needed to tell the mechanisms apart is captured in these three numbers alone.

The Inescapable Cost: When Information is Lost

What happens when we summarize our data with statistics that are not sufficient? We lose information. This is not always a mistake; sometimes it's a necessary compromise.

Consider a microbiologist studying a newly discovered bacterium. They carefully measure its growth rate at various oxygen concentrations and obtain a detailed curve: the bacterium requires a little oxygen, grows best at a low concentration of 2%, and is killed by the 21% oxygen in our atmosphere. The scientist then publishes their finding, labeling the organism a "​​microaerophile​​." This label is a useful, qualitative summary. But look at what has been lost! The label doesn't tell you the optimal oxygen level is 2%. It doesn't tell you how steeply the growth rate falls off, or the precise concentration at which oxygen becomes toxic. The single categorical label is an ​​insufficient statistic​​ for the organism's complex relationship with oxygen. All categorization is a form of information loss.

This trade-off is at the heart of many modern computational methods. In population genetics, the coalescent models that describe the ancestry of genes are so complex that their likelihood function is often intractable. We simply cannot write it down, let alone find a sufficient statistic. In these cases, scientists use ​​Approximate Bayesian Computation (ABC)​​. They knowingly choose a set of plausible but insufficient summary statistics (like genetic diversity, π\piπ, or population differentiation, FSTF_{ST}FST​). They then simulate data under different parameter values and accept the parameter values that produce summary statistics "close" to the ones from the real data. The result is an approximation of the posterior distribution—an estimate of the truth, filtered through the imperfect lens of the chosen summary statistics [@problem_synthesis:2521316]. It's a pragmatic admission that sometimes, a partial answer is better than no answer at all.

Two Sides of a Coin: Information About What?

We end on a subtle but illuminating point. Is all information the same? Let's return to our sample of nnn measurements, X1,X2,…,XnX_1, X_2, \ldots, X_nX1​,X2​,…,Xn​, drawn from a distribution with a parameter θ\thetaθ (say, the mean). Now, let's sort these measurements to create the ​​order statistics​​, Y1≤Y2≤…≤YnY_1 \le Y_2 \le \ldots \le Y_nY1​≤Y2​≤…≤Yn​.

A key result in statistics is that the ​​Fisher information​​, which quantifies how much information the data contains about the parameter θ\thetaθ, is exactly the same for the original sample and the sorted sample. This makes perfect sense. If you are trying to estimate the mean height of a population, a sample of "170 cm, 180 cm" tells you just as much as a sample of "180 cm, 170 cm". The order is irrelevant for an independent sample, so it contains no information about θ\thetaθ.

But something else has changed. The ​​differential entropy​​, which measures the total randomness or "surprise" in the data, has decreased. The sorted vector is less random than the original unordered vector. In fact, for any of the n!n!n! possible permutations of the original data, we get the exact same sorted vector. By sorting, we have discarded the information about which of those n!n!n! permutations we started with. The reduction in entropy is precisely ln⁡(n!)\ln(n!)ln(n!).

This reveals a profound distinction. There is "statistical information" in the Fisher sense—information about a parameter—and there is "information" in the Shannon sense—a measure of the complexity or uncertainty of the data itself. The quest for a sufficient statistic is the art of preserving the former while discarding the latter.

This is the essence of scientific modeling. We look at the world, a system of unfathomable complexity, and we try to find the simple summaries that are sufficient for the questions we are asking. We build a model of planetary motion that cares about mass and velocity but ignores color and composition. We build a model of a gas that cares about temperature and pressure but ignores the trajectory of any single molecule. We search for the simple, elegant clues that point to the underlying truth, and we bravely accept that to see the pattern, we must be willing to ignore the chaos. The search for statistical information is the search for the principles that govern the world.

Applications and Interdisciplinary Connections

We have spent some time exploring the principles and machinery of statistical information, but what is it all for? Is it merely a collection of abstract mathematical ideas? Not at all! The real beauty of a powerful scientific concept lies not just in its internal elegance, but in its "unreasonable effectiveness" in describing the world around us. In this chapter, we will take a journey through a landscape of seemingly disconnected fields—from the intricate dance of molecules in our cells to the deep history of our species and the very fabric of matter—and discover how the thread of statistical information weaves them all together. We will see how learning to read the patterns in data is akin to learning a new language, one that allows us to understand the past, predict the future, and perceive hidden realities.

Decoding the Blueprints of Life: Statistics in Genomics and Bioinformatics

The genome is often called the "book of life," but it's a book written in a four-letter alphabet, billions of characters long. Reading it is one thing; understanding it is another entirely. This is where statistical information becomes our indispensable guide.

Perhaps the most direct and exciting application is in the burgeoning field of personalized medicine. Imagine you could take a person's genetic data and, by cross-referencing it with a vast library of scientific knowledge, compute a single score that predicts their risk for a certain disease or their likely response to a drug. This isn't science fiction; it's the reality of the ​​Polygenic Risk Score (PRS)​​. Large-scale studies, called Genome-Wide Association Studies (GWAS), sift through the genomes of hundreds of thousands of people to find tiny variations associated with a trait. The output of such a study is a massive table of summary statistics—for each genetic variant, it tells us which version tends to increase the trait (the "effect allele") and by how much (the "effect size," or β\betaβ). This table is pure statistical information. To calculate a person's PRS, we simply walk through their genome, and for each relevant variant, we check how many copies of the effect allele they have (0, 1, or 2) and multiply that by the corresponding effect size. Summing these contributions up gives us a personal, statistically-informed prediction. It is a remarkably simple idea with profound implications, turning population-level statistics into individual-level insight.

But this process is fraught with peril. What if the statistical information we are using is itself flawed? A common problem in large genetic studies is that if you unknowingly mix samples from populations with different ancestries, you can create spurious associations. A variant might appear to be linked to a disease simply because it's more common in a group that, for other environmental or genetic reasons, also has a higher rate of that disease. This is called "population stratification," and it can lead to an inflation of false-positive results. How do we fix this? With more statistics!

The brilliant insight of ​​genomic control​​ is to recognize that under the null hypothesis (that there is no true association), the test statistics from a GWAS should follow a known theoretical probability distribution (specifically, a χ2\chi^2χ2 distribution). If we look at the observed distribution of our millions of test statistics, and we see that its median is higher than the theoretical median, it's a good sign that our statistics are systematically inflated. We can calculate an "inflation factor," λ\lambdaλ, which is simply the ratio of the observed median to the expected median. By dividing all of our test statistics by this factor λ\lambdaλ, we can correct for the bias and bring our results back in line with reality. This is a beautiful act of statistical self-correction, using information about the overall distribution of results to improve the reliability of each individual result.

This theme of self-correction appears again in the workhorse tool of bioinformatics, ​​BLAST​​, which searches for similar sequences in massive databases. When you search with a protein sequence, BLAST finds potential matches and assigns them a statistical score (an EEE-value) that tells you how likely it is you'd see a match that good by pure chance. The theory behind these statistics, however, assumes a "typical" amino acid composition. But what if your protein, and a completely unrelated protein in the database, both happen to be rich in, say, proline? Their alignment score might be artificially high simply due to this shared compositional bias, leading to a misleadingly significant EEE-value. This is a common source of false positives. The solution, known as ​​composition-based statistics​​, is to adjust the statistical model on the fly. Instead of using one-size-fits-all statistical parameters, the algorithm looks at the specific compositions of the two sequences being compared and re-calculates the parameters accordingly. This has the effect of "demoting" matches that are only high-scoring because of compositional artifacts, while leaving the scores of true relatives, which have a normal composition, largely unchanged. It is another wonderful example of using statistical information to distinguish true signal from systematic noise.

The sophistication of our models has also evolved. Early attempts to predict the three-dimensional structure of a protein from its amino acid sequence, like the Chou-Fasman method, relied on simple statistical propensities. They asked, "How often is the amino acid Alanine found in an alpha-helix?" This is context-free information. A breakthrough came with methods like the GOR method, which realized that the fate of an amino acid is profoundly influenced by its neighbors. This method calculates the conditional probability of a residue's structure given the identities of the amino acids in a window around it. It uses context-dependent information. This shift from simple frequencies to conditional, context-aware probabilities represents a fundamental leap in how we leverage statistical information, a leap that has been replayed in countless fields of science.

Reading History in Our Genes: The Statistical Archaeologist

The patterns of genetic variation within and between populations are not random. They are an echo of the past, a living record of migrations, expansions, bottlenecks, and adaptations. A population geneticist is a kind of statistical archaeologist, using carefully designed summary statistics as their tools to excavate this history.

Imagine a population of lizards isolated on a mountain that experienced a severe population crash during the last ice age, with only a few individuals surviving. In the aftermath, the population recovered and grew. How would this "bottleneck" event leave its mark on the lizards' DNA today? During the bottleneck, most genetic lineages are lost by chance. The few that survive give rise to the entire modern population. This means that the genealogy of the sampled genes will have a particular shape: long internal branches stretching back to the few ancient survivors, and relatively short terminal branches. Mutations on the long internal branches have had plenty of time to drift to intermediate frequencies, while the short terminal branches mean there will be a deficit of very rare variants (or "singletons").

Population geneticists have designed statistics to detect exactly this kind of pattern. ​​Tajima's DDD​​, for example, compares two different estimates of genetic diversity: one more sensitive to intermediate-frequency variants (π\piπ) and one more sensitive to the total number of variants (SSS). ​​Fu and Li's DDD​​ directly compares the number of singleton variants to the total number of variants. For our post-bottleneck lizards, we would expect to see a deficit of singletons and an excess of intermediate-frequency variants, causing both of these statistics to have tell-tale positive values. By measuring these summary statistics, we can infer the shape of the hidden genealogy and, from that, the dramatic history of the population.

We can take this a step further to tell even more complex stories, such as that of ​​adaptive introgression​​. This occurs when two different populations interbreed, and a gene variant from one population proves to be beneficial in the genetic background of the other and is thus favored by natural selection. To find such a region, we need to find two distinct statistical signatures at once: first, the DNA in that region must show a high probability of originating from the donor population (a local ancestry signal), and second, it must show the hallmarks of a recent selective sweep, such as an unusually long, un-broken haplotype that has risen to high frequency (a haplotype homozygosity signal). A powerful statistical test will not just look for these signals in isolation but will combine them in a rigorous, composite framework. The most sophisticated methods will define ancestry-specific groups of haplotypes within the admixed population and compare them, all while carefully building a null model based on simulations of the population's specific demographic history and local recombination rates to avoid being fooled by confounders. This is the pinnacle of statistical archaeology: weaving together multiple threads of evidence to reconstruct a specific, detailed story of evolution in action.

When the Math is Too Hard: The Dawn of Algorithmic Inference

What happens when the processes we want to study are so complex that we can no longer write down a simple equation for the probability of our data? The history of a population, with all its randomness of mating, mutation, and migration, is a perfect example. We can easily write a computer program to simulate this process, but we often can't write down a neat mathematical likelihood function. Does this mean we must give up on statistical inference?

Absolutely not. We turn to a clever and powerful idea called ​​Approximate Bayesian Computation (ABC)​​. The intuition is wonderfully simple: if I can't calculate the probability of my observed data, I'll instead create thousands of simulated "universes" on my computer. In each simulation, I'll pick a set of parameters (like population size or selection strength) from some prior distribution. I'll let the simulation run, and then I'll compute a set of summary statistics from my simulated data. I then compare the summary statistics from each simulation to the ones from my actual, observed data. If they are close enough, I "accept" the parameters that went into that simulation. The collection of all accepted parameters forms an approximation of the posterior distribution—it's the set of parameters that are capable of producing a world that looks like ours.

This workflow allows us to tackle incredibly complex problems. For instance, if we want to infer the strength and timing of a selective sweep, we can simulate sweeps under many different scenarios and find the ones that best reproduce the observed pattern of haplotype homozygosity, linkage disequilibrium, and the site frequency spectrum. But the success of this whole enterprise hinges on a crucial choice: what summary statistics should we use? This is the art of ABC. The statistics must be "sufficient" enough to capture the information that distinguishes our competing hypotheses. When trying to distinguish a population expansion from a bottleneck or long-term structure, we can't just pool all our data. We must use statistics that are sensitive to the differences between populations, like the ​​joint site frequency spectrum (jSFS)​​, which tabulates how alleles are shared across populations, and the fixation index (FSTF_{ST}FST​). Furthermore, we must choose statistics that are robust to the limitations of our real data, such as being unphased or having uncertain ancestral states, which is why using a "folded" SFS is often wise. ABC, therefore, is not just a brute-force technique; it is a thoughtful marriage of simulation power and statistical insight.

The Deep Unity: From Patterns in Crystals to the Fabric of Matter

The reach of statistical information extends far beyond the realm of biology and into the heart of the physical world. Consider the problem of determining the structure of a crystal using neutron diffraction. When a beam of neutrons passes through a crystal, it scatters off the atomic nuclei, creating a complex diffraction pattern of bright spots. This pattern seems almost random in its intensities, but it contains profound information about the arrangement of atoms.

One of the most fundamental properties of a crystal is whether it possesses a center of symmetry (centrosymmetric) or not (non-centrosymmetric). A.J.C. Wilson showed that one could determine this property simply by looking at the statistics of the diffraction intensities. The structure factor, F(h)F(\mathbf{h})F(h), which determines the intensity of a diffraction spot, is a sum over the contributions of all atoms in the unit cell. If the number of atoms is large, the Central Limit Theorem tells us that F(h)F(\mathbf{h})F(h) will behave like a random variable with a Gaussian distribution. Critically, if the crystal is centrosymmetric, F(h)F(\mathbf{h})F(h) is purely real. If it's non-centrosymmetric, it's a complex number with independent real and imaginary parts. This seemingly small difference leads to completely different probability distributions for the normalized intensities, z=∣E∣2z = |E|^2z=∣E∣2. By measuring the second moment of the observed intensities, z2‾\overline{z^2}z2, and comparing it to the theoretical predictions (a value of 3 for the centric case, and 2 for the acentric case), we can make a robust inference about the crystal's hidden symmetry. It's a masterful piece of reasoning: a fundamental, non-random property of the crystal is revealed by the statistical distribution of what appears to be random noise.

This journey culminates in perhaps the most surprising and beautiful connection of all, linking the abstract world of statistics to the quantum mechanics of matter. In statistics, a concept called ​​Fisher Information​​, IFI_FIF​, quantifies how much information a random variable carries about a parameter. For a probability distribution p(r)p(\mathbf{r})p(r), the Fisher information density is proportional to ∣∇p(r)∣2/p(r)|\nabla p(\mathbf{r})|^2/p(\mathbf{r})∣∇p(r)∣2/p(r). This is a purely statistical concept.

Now, let's step into the world of quantum chemistry. A key goal of Density Functional Theory (DFT) is to approximate the kinetic energy of a system of electrons. One of the fundamental components of this energy is the von Weizsäcker kinetic energy density, τW(r)\tau_W(\mathbf{r})τW​(r), a quantity derived from the electron density n(r)n(\mathbf{r})n(r). Its formula is τW(r)=∣∇n(r)∣2/(8n(r))\tau_W(\mathbf{r}) = |\nabla n(\mathbf{r})|^2 / (8 n(\mathbf{r}))τW​(r)=∣∇n(r)∣2/(8n(r)). If we simply consider the electron density of an NNN-electron system as being proportional to a probability distribution, n(r)=Np(r)n(\mathbf{r}) = N p(\mathbf{r})n(r)=Np(r), a remarkable identity emerges through trivial substitution: the physical quantity TW[n]T_W[n]TW​[n] is directly proportional to the statistical quantity IF[p]I_F[p]IF​[p], with TW[n]=N8IF[p]T_W[n] = \frac{N}{8} I_F[p]TW​[n]=8N​IF​[p].

Stop and marvel at this. A quantity from quantum physics that describes the energy of motion of electrons is, up to a constant, the very same mathematical object that a statistician uses to describe the information content of a distribution. This deep unity reveals that the curvature and gradients in the electron cloud, which give rise to kinetic energy, are synonymous with the "information" embedded in that cloud's shape. It is in discovering such unexpected bridges between disparate domains that we glimpse the true, underlying coherence of the natural world—a world written in the language of statistical information.