try ai
Popular Science
Edit
Share
Feedback
  • A Brief History of Information Theory: From Bits to Biology and Beyond

A Brief History of Information Theory: From Bits to Biology and Beyond

SciencePediaSciencePedia
Key Takeaways
  • Information is fundamentally a measure of the reduction of uncertainty, quantified by the "bit," a concept that launched the digital age.
  • Thermodynamic entropy and Shannon's information entropy are physically equivalent, revealing that information is a tangible property of the universe.
  • The instructions for life are not solely in DNA; heritable information is also stored epigenetically and emerges from interactions with the environment.
  • The principles of information theory act as a unifying framework, providing essential tools to understand systems as diverse as biological ecosystems, quantum computers, and the causal structure of spacetime.

Introduction

What do a telegraph message, a strand of DNA, and the laws of thermodynamics have in common? The answer lies in one of the 20th century's most profound and far-reaching ideas: information. Born from the practical problem of sending clear messages over noisy wires, information theory has evolved into a fundamental language for science itself. It addresses a knowledge gap that spans disciplines, revealing that the same mathematical principles governing data transmission also describe the workings of life, the behavior of matter, and the very fabric of reality.

This article explores the remarkable journey of this concept. We will trace its history and its stunning power to connect seemingly unrelated worlds. The first section, ​​Principles and Mechanisms​​, delves into the core ideas. We will uncover what information truly is, explore its deep and surprising identity with physical entropy, and see how this particulate, probabilistic way of thinking revolutionized our understanding of heredity and development. Following this, the section on ​​Applications and Interdisciplinary Connections​​ will showcase this theory in action. We will see how the language of bits and channels provides a powerful lens for biologists, physicists, and computer scientists, offering insights into everything from the diversity of ecosystems and the engineering of living cells to the ultimate limits of computation and the bizarre logic of time travel. Prepare to see the world not as a collection of things, but as a tapestry woven from information.

Principles and Mechanisms

What Is Information, Really?

Let us begin with a question that seems deceptively simple: what is ​​information​​? Your first instinct might be to think about meaning, about the content of a sentence or the plot of a story. But in a scientific sense, this is not where the story begins. The revolution that created information theory began by stripping away the semantics and focusing on something far more fundamental: the reduction of uncertainty.

Imagine you are an engineer in the 1930s, tasked with sending messages over a wire. You devise a clever system. For each symbol you want to send, you divide a small window of time into 8 distinct slots. You then send a single electrical pulse in just one of those slots. The receiver on the other end simply has to figure out, "In which of the 8 slots did the pulse arrive?" Before the pulse is sent, there are 8 possibilities. Once it is received, there is only one. Information, in this starkly beautiful view, is that which resolves the uncertainty.

How much information have you conveyed? Well, how many "yes or no" questions would you need to ask to pinpoint the correct slot? You could ask, "Was it in the first four slots?" If the answer is yes, you ask, "Was it in the first two?" and so on. With three well-posed questions, you can always find the answer. So, we say the transmission of one symbol conveys 3 ​​bits​​ of information. The "bit" is the fundamental atom of information, the answer to a single yes-or-no question. If you have MMM equally likely possibilities, the amount of information required to specify one of them is log⁡2(M)\log_{2}(M)log2​(M). For our 8 slots, log⁡2(8)=3\log_{2}(8) = 3log2​(8)=3. This simple, powerful idea, formalized by Claude Shannon, launched the digital age. It's an idea that cared not for what you were saying, only for the number of possibilities you were choosing from.

The Physicist's Demon and the Price of Knowledge

For a time, this concept of information seemed to belong to the world of engineers and mathematicians. But a deep and startling connection was lurking in the realm of physics, one that would elevate information from a measure of messages to a fundamental property of the universe itself. The connection lies in a concept that physicists had been wrestling with for decades: ​​entropy​​.

In thermodynamics, entropy is often described, rather vaguely, as a measure of "disorder." But Ludwig Boltzmann gave it a much more precise meaning. Imagine a box containing a single gas molecule. We can conceptually divide this box into, say, N=210=1024N = 2^{10} = 1024N=210=1024 tiny, equal-sized cells. If the molecule could be in any one of these cells with equal probability, the entropy of the system is given by Boltzmann's famous formula, S=kBln⁡(W)S = k_B \ln(W)S=kB​ln(W), where WWW is the number of possible microscopic arrangements (microstates)—in our case, W=1024W = 1024W=1024.

Now, look at this from an information perspective. If I know the molecule is in the box, but I don't know which of the 1024 cells it occupies, how much information am I missing? Using the very same logic as our telegraph problem, the missing information is I=log⁡2(1024)=10I = \log_2(1024) = 10I=log2​(1024)=10 bits. The astonishing insight, a cornerstone of modern physics, is that these two quantities are not just analogous; they are, up to a constant factor, the same thing. The thermodynamic entropy SSS is simply the Shannon information III you lack about the system's exact microstate, scaled by a physical constant: S=I⋅(kBln⁡2)S = I \cdot (k_B \ln 2)S=I⋅(kB​ln2).

This is a profound unification. It tells us that information is physical. Our ignorance about the world has a measurable, physical consequence. To decrease the entropy of a system—to clean your room, for example—you must increase your information about it. And conversely, to gain information—to measure which cell the particle is in—requires a physical interaction that, as it turns out, has a minimum thermodynamic cost. There is no such thing as a free bit.

The Atoms of Heredity

This new way of thinking—seeing the world in terms of discrete possibilities and probabilities—was not confined to physics. It was a wave that was about to crash into the shores of biology. Around 1900, the long-neglected work of Gregor Mendel was rediscovered, and this time, the world was ready. Why? Yes, microscopes had improved, and biologists were beginning to see chromosomes inside cells. But a deeper intellectual shift had occurred, one imported directly from physics.

For decades, physicists had been explaining the smooth, continuous properties of gases—their temperature, their pressure—as the collective statistical behavior of vast numbers of discrete, jittering atoms. No one could see a single atom, but the atomic theory, powered by ​​statistical mechanics​​, explained the macroscopic world with stunning success. Scientists had learned to think in terms of hidden, particulate units whose probabilistic behavior gives rise to the phenomena we observe.

This was precisely the mental framework needed to grasp Mendel's genius. The prevailing theory of heredity had been one of "blending," as if traits were like paints mixing together. Mendel proposed something radically different: heredity was particulate. Traits were governed by discrete "factors" (which we now call genes) that are passed down, shuffled, but never blended. An offspring inherits one factor for a trait from each parent, and the combination of these discrete units determines the observable characteristic. It was a probabilistic, quantitative, and particulate theory of heredity. It was, in essence, an information theory of biology. Mendel’s factors were the "atoms" of inheritance, and his laws were the statistical mechanics that governed their transmission.

The Recipe Is Not the Cake

The discovery of genes and later DNA as the "atoms" of heredity led to a new temptation. If the entire plan for an organism is written in the DNA sequence, is development just the process of a pre-formed blueprint simply growing larger? This is a modern version of an old debate: ​​preformation versus epigenesis​​. Preformationism is the idea that a tiny, complete organism—a homunculus—is curled up in the sperm or egg, and development is just inflation. Epigenesis, on the other hand, argues that complexity arises progressively from a simpler state through a series of interactions.

Modern biology provides stunning vindication for epigenesis, showing that the genetic code is less like a blueprint and more like a recipe. A recipe is a set of instructions, but it requires a kitchen, ingredients, and a chef to interpret and act upon them. Consider the development of our own gut. A mammal raised in a completely sterile, germ-free environment will have a stunted and dysfunctional intestinal tract. The intricate, finger-like villi that absorb nutrients will be underdeveloped, and a crucial part of the immune system will fail to mature.

Why? Because the host's DNA alone doesn't have all the "information" needed to build a proper gut. It relies on a constant dialogue with the trillions of bacteria that make up our microbiome. These microbes, external biotic factors, send chemical signals that tell the host's cells how to divide, differentiate, and organize. The final, complex architecture of the gut is not pre-formed in the embryo; it emerges from the interaction between the genetic recipe and its environment. The information required for development is not contained solely within the genome.

Echoes of the Past: Information Beyond DNA

The story gets even stranger and more wonderful. We tend to think that the information passed from parent to child is locked in the sequence of As, Ts, Cs, and Gs in our DNA. But what if the way the recipe book is annotated—the sticky notes, the highlights, the folded corners—could also be passed down? This is the revolutionary field of ​​transgenerational epigenetic inheritance​​.

Imagine a thought experiment where male mice are exposed to a chemical that doesn't change their DNA sequence at all, but attaches specific chemical tags (like acetylation marks) to the proteins that package the DNA in their sperm. These tags act like instructions, perhaps telling a gene to be "read more loudly" or "be quiet." Remarkably, when these mice have offspring, and even grand-offspring, that were never exposed to the chemical, they can inherit the behavioral traits caused by it. The chemical tags, the annotations on the DNA, have been passed down through the generations.

This is a form of heritable information that exists "on top of" the genetic code. It shows that an organism's developmental trajectory is guided by layers of information, some of which are written in the ink of DNA, and others written in a more transient, but still heritable, chalk. This is epigenesis in its most profound form: development is not the execution of a static program but a dynamic process guided by a rich, multi-layered information system that is responsive to the environment.

The Signature of Life

From telegraphs to thermodynamics, from pea plants to our own development, the concept of information has proven to be a unifying thread of incredible power. This brings us to perhaps the most fundamental question of all: What is ​​life​​? Could it be that the essence of life, too, is best understood in the language of information?

When we design missions to search for life on other worlds, we are forced to confront this question head-on. What do we even look for? Different definitions of life lead to vastly different search strategies.

  • A ​​"metabolism-first"​​ view suggests life is fundamentally a self-sustaining chemical engine. We should look for signs of this engine running: chemical cycles far from equilibrium, energy being consumed, and characteristic waste products. The information-storage part, like DNA, might be a later addition.

  • An ​​"information-first"​​ view (or "genetics-first") posits that life begins with a replicator—a molecule, like RNA, that can store information and make copies of itself. The priority here is to find long, complex polymers with non-random sequences, evidence of a code. The metabolic engine could be co-opted later.

The working definition used by NASA beautifully marries these two ideas: life is "a self-sustaining chemical system capable of Darwinian evolution." This definition has two parts. The "self-sustaining chemical system" is the metabolic engine, the hardware that keeps itself running by processing energy and matter. The "capable of Darwinian evolution" part is all about information. It requires a system of heredity (a code), variation (mutations in that code), and selection. Life, in this view, is not just an engine; it is an engine that can learn. It is an information-processing system that uses energy to preserve and propagate the information that allows it to exist.

And so, our journey comes full circle. The simple, practical question of how to send a message down a wire led to a concept so fundamental that it connects the statistical flutter of atoms to the grand tapestry of heredity and evolution. The bit, that humble unit of choice, has become a key part of the language we use to ask the most profound question of all: Are we alone in the universe? When we search for life, we are no longer just looking for strange chemistries; we are looking for a signature, the signature of information at work.

Applications and Interdisciplinary Connections

We have spent some time exploring the beautiful mathematical structure that Claude Shannon erected—a theory of information. We've seen its parts: entropy as a measure of surprise, mutual information as a measure of shared surprise, and channel capacity as the ultimate speed limit for communication. But a theory, no matter how elegant, is just a skeleton. To see its true power, we must see it in the flesh, to watch it walk and breathe in worlds far beyond the telephone lines and telegraphs for which it was first conceived. You might be surprised to find that this language of bits and noise is spoken fluently by biologists, physicists, computer scientists, and even cosmologists. It seems that nature, in its endless complexity, has been playing by these rules all along.

So, let's go on a little tour. We will not be engineers for a day, but explorers. We'll see how these ideas provide a new lens through which to view the world, from the grand tapestry of life on Earth down to the very fabric of spacetime.

Information as the Currency of Life

If you look at the natural world, what do you see? A relentless, chaotic struggle for existence. But look closer, and you see something else: a world awash in information. A flower's color is a signal to a bee. A predator's scent is a warning to its prey. And deep within every cell, a torrent of information flows, dictating its form and function. It is no surprise, then, that Shannon’s tools have become indispensable for understanding biology.

Let's start with a walk in the woods. You see a great variety of plants and animals. How do you quantify this "biodiversity"? You could just count the number of species, but that feels incomplete. An ecosystem with two species in equal number feels more diverse than one where 99% of the individuals belong to one species and only 1% to the other. The ecologist's problem is to find a number that captures this feeling. It turns out that if you write down a few simple, reasonable axioms for what a diversity measure should do—for instance, that it should be continuous and treat all species symmetrically—you are led, almost by magic, to a single mathematical form: the Shannon entropy, H=−∑ipiln⁡piH = -\sum_{i} p_i \ln p_iH=−∑i​pi​lnpi​. Here, pip_ipi​ is just the proportion of the iii-th species. The uncertainty in picking a random individual from the ecosystem is its diversity. This measure has a wonderful property: it is particularly sensitive to the presence of rare species. In a highly uneven community, the surprise contribution, −piln⁡pi-p_i \ln p_i−pi​lnpi​, from a very rare species can be much larger than its tiny abundance pip_ipi​ would suggest, a mathematical acknowledgment that these rare members are crucial to the richness of the whole.

From the forest, let’s zoom down into the nucleus of a single cell. Here, long strands of DNA are spooled and packed, decorated with chemical tags called histone modifications. For a long time, biologists knew these tags were associated with whether a gene was turned "on" or "off," but the relationship was fuzzy. It's a classic signaling problem: how much does the presence of a histone mark "tell" us about the expression of a nearby gene? By collecting vast amounts of data, we can build a contingency table counting how often a gene is active when a mark is present, inactive when it's present, and so on. From this table, we can compute the probabilities and then, straight from the textbook, calculate the mutual information I(G;H)I(G; H)I(G;H) between gene expression GGG and the histone mark HHH. This gives us a single number, in bits, that quantifies the strength of the regulatory connection, filtering out the noise and revealing the underlying logic of the cell's control system.

This is so powerful that biologists are no longer content just to observe. They have become engineers. In the field of synthetic biology, scientists build new gene circuits inside cells, much like an electrical engineer builds circuits with resistors and capacitors. Imagine designing a circuit where a cell should produce a fluorescent protein (the output) in response to a chemical inducer (the input). The cell is a noisy environment; the process isn't perfect. How well does our designed circuit work? We can model the entire cell as a communication channel. The mutual information between the inducer concentration and the fluorescence level tells us exactly how many distinct input levels the cell can reliably distinguish. We can then ask, what is the best this circuit can do? By optimizing over all possible ways of presenting the input signal, we can calculate the channel capacity of our gene circuit, just as Shannon did for a telephone wire. This isn't just an academic exercise. In designing CAR-T cells for cancer immunotherapy, we are programming an immune cell to "read" the antigen density on other cells and "decide" if they are cancerous. Maximizing the channel capacity of this recognition process means designing a more effective, more discriminating, and safer therapy. Shannon's theory is literally becoming a matter of life and death.

Information in the Physical and Computational World

The notion of information having a physical home is a deep one. Think about a piece of clay. If you deform it, it remembers its new shape. That memory—that information about its history of deformation—must be stored in the clay itself. This simple idea has profound consequences in fields like computational mechanics, which simulates the behavior of materials under stress. In methods like the Material Point Method (MPM), the material is represented by a cloud of "particles" that move through a fixed computational grid. The grid is useful for calculating forces and spatial gradients at a single instant, but it is wiped clean and reinitialized at every step of the simulation. All the information about the material's history—its stress, its strain, its past deformations—must be carried by the particles themselves. Why? Because history is a property of the material points, and the particles are our stand-ins for those points. The information follows the matter. It is a beautiful illustration of the distinction between the Eulerian view (standing on the riverbank watching the water flow past) and the Lagrangian view (floating in a boat carried by the current).

This idea of a "state" as a snapshot of information is the absolute bedrock of computer science. The famous Cook-Levin theorem, which established the concept of NP-completeness, is built on this. The proof shows that any problem that can be solved by a non-deterministic Turing machine in a reasonable amount of time can be converted into a giant Boolean satisfiability problem. How? By creating a "tableau," which is nothing more than a giant grid where each row is a complete description of the computer at one moment in time: the state of its processor, the position of its read/write head, and the entire contents of its tape. The entire computation, from start to finish, is laid out as a static object. The rules of the machine are translated into logical clauses connecting one row to the next. The original problem has an answer if, and only if, this giant logical formula is satisfiable. The history of a computation is a piece of information, and its properties determine the fundamental limits of what is computable.

The story gets even stranger when we cross into the quantum world. The quantum analog of Shannon entropy is the von Neumann entropy, S(ρ)=−Tr(ρlog⁡2ρ)S(\rho) = -\text{Tr}(\rho \log_2 \rho)S(ρ)=−Tr(ρlog2​ρ), where ρ\rhoρ is the density matrix describing a quantum system. This quantity, it turns out, gives the ultimate limit for compressing quantum information, a process known as Schumacher compression. But it does more. In fundamental physics, it has become a central tool for understanding the most puzzling feature of quantum mechanics: entanglement. Consider a quantum system in its quiet ground state. If you suddenly change its Hamiltonian (a "quantum quench"), it's like shaking the system violently. Pairs of entangled quasi-particles are created everywhere and fly apart. If you look at a subregion of the system, its entanglement with the rest of the system grows over time. The amount of that entanglement, a measure of how deeply quantum-linked the region is with its surroundings, is given precisely by the von Neumann entropy. In certain systems, its growth follows a beautiful, simple geometric rule that we can calculate. The same number that tells an engineer the limit of data compression tells a physicist how information, in the form of quantum entanglement, spreads through the universe. Isn't that remarkable?

Information at the Edge of Spacetime and Causality

We have seen information in living cells, in ecosystems, in computers, and in quantum fields. The final leg of our journey takes us to the very edge of reality, to the intersection of information theory and Einstein's theory of General Relativity.

General Relativity allows for bizarre spacetime geometries, including the possibility of "Closed Timelike Curves" (CTCs)—paths through spacetime that an observer could follow to return to their own past. This opens a Pandora's box of paradoxes. Consider the "bootstrap paradox": A physicist finds an artifact with the complete theory of wormholes written on it. She uses the theory to build a time machine, travels to the past, and leaves the artifact to be found. So, where did the information—the theory itself—come from? She learned it from the artifact, but she is the one who put it there.

Within the framework of self-consistent CTCs, the answer is as profound as it is strange: the information has no origin. It is a globally consistent solution to the laws of physics along a closed loop in spacetime. The information simply is, a feature of the universe's history, not created at any single point but existing as a self-supporting, acausal loop. The information doesn't need a "beginning" because the timeline it exists on has no beginning. This forces us to abandon our intuitive notion that information must be created before it can be received. In these exotic corners of the cosmos, cause and effect can form a closed circle.

From the pragmatic design of a cancer therapy to the mind-bending logic of a causal loop, the thread that connects them all is the concept of information. Shannon's work gave us more than just a way to build better communication systems. It gave us a quantitative, universal language to describe the order, pattern, and structure hidden in the chaos of the world. It is a lens that, once you learn to see through it, changes how you see everything. And the journey of discovery it began is far, far from over.