try ai
Popular Science
Edit
Share
Feedback
  • The Universal Pythagorean Theorem

The Universal Pythagorean Theorem

SciencePediaSciencePedia
Key Takeaways
  • The Pythagorean theorem is a direct consequence of orthogonality in any space with an inner product, not just a geometric fact about triangles.
  • This principle extends to abstract function spaces, where it appears as Parseval's theorem, equating the total energy of a signal to the sum of the energies of its orthogonal components.
  • In statistics and machine learning, the theorem manifests as orthogonal projection (the method of least squares), which decomposes data into an optimal solution and an orthogonal error.
  • The theorem's structure provides the foundation for probability in quantum mechanics and for measuring "distance" between probability distributions in information theory.

Introduction

The Pythagorean theorem, a2+b2=c2a^2 + b^2 = c^2a2+b2=c2, is one of the most recognized results in mathematics, deeply associated with the geometry of right-angled triangles. However, this familiar equation is merely the visible tip of a conceptual iceberg—a specific instance of a universal principle that weaves through seemingly unrelated fields. The common perception of the theorem as an isolated geometric rule forms a knowledge gap, obscuring its true power as a fundamental principle of orthogonality. This article bridges that gap by revealing the theorem's profound and unifying influence across science.

This article will guide you on a journey to uncover this hidden unity. In the "Principles and Mechanisms" chapter, we will reframe the theorem using the language of vectors and inner products to expose its algebraic heart. We will then witness how this core idea blossoms in the infinite-dimensional world of function spaces, the curved fabric of spacetime, and the abstract realm of information. Following this, the "Applications and Interdisciplinary Connections" chapter will demonstrate how this generalized principle becomes a practical workhorse in data science, quantum mechanics, and signal processing, providing a unified geometric intuition for a vast range of phenomena.

Principles and Mechanisms

It’s one of the first beautiful pieces of mathematics we ever learn: for a right-angled triangle, a2+b2=c2a^2 + b^2 = c^2a2+b2=c2. It seems like a simple, self-contained fact about triangles and squares drawn on a flat piece of paper. But what if I told you that this little equation is a keyhole? And if you peer through it, you’ll see a vast, interconnected landscape that stretches from the strings of a violin to the fabric of spacetime, and even into the very nature of information and uncertainty. The Pythagorean theorem isn’t just a rule; it’s a symptom of a much deeper, more fundamental structure in the universe. Let’s embark on a journey to uncover this hidden unity.

The Secret in the Inner Product

Let's begin by reimagining our triangle. Instead of thinking about side lengths, let's think about vectors—arrows with a length and a direction. The two legs of our right triangle can be seen as two vectors, let’s call them a\mathbf{a}a and b\mathbf{b}b, that are perpendicular to each other. The hypotenuse is then the vector representing their sum, c=a+b\mathbf{c} = \mathbf{a} + \mathbf{b}c=a+b.

In this language, the theorem becomes: if a\mathbf{a}a is perpendicular to b\mathbf{b}b, then ∥a+b∥2=∥a∥2+∥b∥2\|\mathbf{a}+\mathbf{b}\|^2 = \|\mathbf{a}\|^2 + \|\mathbf{b}\|^2∥a+b∥2=∥a∥2+∥b∥2, where ∥v∥\|\mathbf{v}\|∥v∥ means the length of vector v\mathbf{v}v. How does the mathematics know when vectors are perpendicular? It knows through a wonderful operation called the ​​inner product​​ (or dot product in familiar Euclidean space). For two vectors x\mathbf{x}x and y\mathbf{y}y, their inner product is written as ⟨x,y⟩\langle \mathbf{x}, \mathbf{y} \rangle⟨x,y⟩.

This single operation is the heart of the matter. It defines everything we need:

  1. ​​Length:​​ The squared length of a vector x\mathbf{x}x is simply the inner product with itself: ∥x∥2=⟨x,x⟩\|\mathbf{x}\|^2 = \langle \mathbf{x}, \mathbf{x} \rangle∥x∥2=⟨x,x⟩.
  2. ​​Angle:​​ Two non-zero vectors x\mathbf{x}x and y\mathbf{y}y are declared ​​orthogonal​​ (the grown-up word for perpendicular) if and only if their inner product is zero: ⟨x,y⟩=0\langle \mathbf{x}, \mathbf{y} \rangle = 0⟨x,y⟩=0.

Now, let's open up the expression for the squared length of the hypotenuse, ∥a+b∥2\|\mathbf{a}+\mathbf{b}\|^2∥a+b∥2, using the basic algebraic rules of inner products:

∥a+b∥2=⟨a+b,a+b⟩=⟨a,a⟩+⟨a,b⟩+⟨b,a⟩+⟨b,b⟩\|\mathbf{a}+\mathbf{b}\|^2 = \langle \mathbf{a}+\mathbf{b}, \mathbf{a}+\mathbf{b} \rangle = \langle \mathbf{a}, \mathbf{a} \rangle + \langle \mathbf{a}, \mathbf{b} \rangle + \langle \mathbf{b}, \mathbf{a} \rangle + \langle \mathbf{b}, \mathbf{b} \rangle∥a+b∥2=⟨a+b,a+b⟩=⟨a,a⟩+⟨a,b⟩+⟨b,a⟩+⟨b,b⟩

In any reasonable space (a real inner product space), the order doesn't matter, so ⟨a,b⟩=⟨b,a⟩\langle \mathbf{a}, \mathbf{b} \rangle = \langle \mathbf{b}, \mathbf{a} \rangle⟨a,b⟩=⟨b,a⟩. This simplifies to:

∥a+b∥2=∥a∥2+∥b∥2+2⟨a,b⟩\|\mathbf{a}+\mathbf{b}\|^2 = \|\mathbf{a}\|^2 + \|\mathbf{b}\|^2 + 2\langle \mathbf{a}, \mathbf{b} \rangle∥a+b∥2=∥a∥2+∥b∥2+2⟨a,b⟩

Look at that! The familiar Pythagorean theorem, ∥a+b∥2=∥a∥2+∥b∥2\|\mathbf{a}+\mathbf{b}\|^2 = \|\mathbf{a}\|^2 + \|\mathbf{b}\|^2∥a+b∥2=∥a∥2+∥b∥2, holds if and only if that final term, 2⟨a,b⟩2\langle \mathbf{a}, \mathbf{b} \rangle2⟨a,b⟩, vanishes. And it vanishes precisely when ⟨a,b⟩=0\langle \mathbf{a}, \mathbf{b} \rangle = 0⟨a,b⟩=0—the definition of orthogonality!

This isn't just a proof; it's a revelation. The Pythagorean theorem is not a geometric coincidence. It is the most direct consequence of defining length and angle through an inner product. Any space, no matter how abstract, that has a consistent notion of an inner product will automatically have a Pythagorean theorem baked into its very fabric. This equivalence is the key that unlocks all the generalizations to come. This same inner product algebra also gives us other geometric gems for free, like the ​​Parallelogram Law​​, which relates the lengths of a parallelogram's sides to its diagonals: ∥u+v∥2+∥u−v∥2=2(∥u∥2+∥v∥2)\|\mathbf{u}+\mathbf{v}\|^2 + \|\mathbf{u}-\mathbf{v}\|^2 = 2(\|\mathbf{u}\|^2 + \|\mathbf{v}\|^2)∥u+v∥2+∥u−v∥2=2(∥u∥2+∥v∥2).

An Orchestra of Orthogonal Functions

So, any space with an inner product has a Pythagorean theorem. But what kinds of spaces have inner products? We are used to spaces of arrows in 2D or 3D. But what about... functions? Can we treat a function, like f(t)=sin⁡(t)f(t) = \sin(t)f(t)=sin(t), as a "vector"?

Absolutely! Think about it: a vector in 3D is a list of three numbers (x,y,z)(x, y, z)(x,y,z). A function f(t)f(t)f(t) is like a vector with an infinite number of components, one for each value of ttt. We can define an inner product for two real functions, f(t)f(t)f(t) and g(t)g(t)g(t), on an interval, say from −1-1−1 to 111, by using an integral:

⟨f,g⟩=∫−11f(t)g(t) dt\langle f, g \rangle = \int_{-1}^{1} f(t)g(t) \, dt⟨f,g⟩=∫−11​f(t)g(t)dt

This integral acts just like the dot product: it takes two functions and spits out a single number. It obeys all the right rules, and from it, we can define the "length" of a function (its ​​norm​​) and what it means for two functions to be "orthogonal".

This is where things get really interesting. Consider the fundamental building blocks of sound and signals: sine and cosine waves. It turns out that functions like sin⁡(t)\sin(t)sin(t), sin⁡(2t)\sin(2t)sin(2t), sin⁡(3t)\sin(3t)sin(3t),... form an ​​orthogonal set​​. A musical chord is a sum of these fundamental notes. A complex signal from a radio telescope is a sum of simple electromagnetic waves. In the language of function spaces, a complex signal S(t)S(t)S(t) is a vector sum of orthogonal function-vectors.

What does the Pythagorean theorem tell us here? It tells us that the total power of the signal (its squared norm) is simply the sum of the powers of its individual components! If a signal is made of three orthogonal frequencies with amplitudes C1,C2,C3C_1, C_2, C_3C1​,C2​,C3​, its total power is simply ∣C1∣2+∣C2∣2+∣C3∣2|C_1|^2 + |C_2|^2 + |C_3|^2∣C1​∣2+∣C2​∣2+∣C3​∣2. This is a profoundly important result in physics and engineering, and it's just Pythagoras in a new outfit.

This idea reaches its zenith in ​​Parseval's Theorem​​ for Fourier series. It states that the total "energy" of a function, given by ∫∣f(x)∣2dx\int |f(x)|^2 dx∫∣f(x)∣2dx, is exactly equal to the sum of the squares of its coordinates in the basis of sine and cosine functions. This is the Pythagorean theorem for an infinite-dimensional space of functions. A vector's length doesn't care if you measure its components in one orthonormal basis or another; the sum of the squares always comes out the same. Parseval's theorem is the ultimate expression of this invariance, extending it from finite-dimensional arrows to the infinite world of functions.

A Deeper Geometry

With our powerful new perspective, let's look back at geometry itself. We've taken the theorem out of geometry and into algebra; now let's bring it back, and see how it reshapes our understanding of space.

The View from Curved Space

On the curved surface of the Earth, a large triangle with vertices at the North Pole, a point on the equator in Africa, and a point on the equator in South America can have three right angles. The familiar a2+b2=c2a^2+b^2 = c^2a2+b2=c2 fails spectacularly.

However, if you look at a tiny, microscopic patch of the Earth's surface, it looks pretty flat. For an infinitesimal right triangle, the theorem still holds. This is the central idea of differential geometry, the language of Einstein's General Relativity. In a curved space or spacetime, the Pythagorean theorem becomes the local, infinitesimal rule for measuring distance. The formula for the infinitesimal distance dsdsds, called the ​​line element​​, generalizes from ds2=dx2+dy2ds^2 = dx^2 + dy^2ds2=dx2+dy2 in a flat plane to:

ds2=∑i,jgijdxidxjds^2 = \sum_{i,j} g_{ij} dx^i dx^jds2=i,j∑​gij​dxidxj

The object gijg_{ij}gij​ is the famous ​​metric tensor​​, and you can think of it as a set of local correction factors that tell you how the Pythagorean theorem works in a specific, possibly warped, coordinate system. Even in the bizarre, non-Euclidean geometry of a material like "hyperbolene," where distances are stretched and distorted, the length of any path is found by adding up (integrating) these infinitesimal Pythagorean distances. The theorem is no longer a global truth, but it becomes the fundamental local law of all geometry.

More Dimensions, More Faces

The theorem also generalizes in another, stunningly beautiful way in flat space. Imagine a tetrahedron in 3D, with one corner at the origin and the other three vertices lying on the x, y, and z axes. This "right-angled" tetrahedron has four faces. Three of them are right triangles in the coordinate planes—we can call them the "leg" faces. The fourth face, spanning the three vertices on the axes, is the "hypotenuse" face.

​​De Gua's theorem​​, a 3D analogue of Pythagoras's, states that the square of the area of the hypotenuse face is equal to the sum of the squares of the areas of the three leg faces. This is not a coincidence! This pattern holds true in any number of dimensions, where the squared (n−1)(n-1)(n−1)-dimensional hypervolume of the "hypotenuse" facet of a right-angled n-simplex is the sum of the squared hypervolumes of the other nnn "leg" facets. It’s the same Pythagorean melody, played on an instrument of higher-dimensional volumes.

The Geometry of Information

We've seen the theorem in spaces of arrows, functions, and even spacetime itself. Could we possibly push it further? What if the "points" in our space were not locations, but abstract concepts, like probability distributions?

Welcome to the mind-bending field of ​​Information Geometry​​. In this world, every point is a probability distribution. One point might be the perfect bell curve of a Gaussian distribution; another might be the distribution of possible outcomes for a weighted die.

How do you measure the "distance" between two beliefs? There's no physical ruler. Instead, information theory provides a tool: the ​​Kullback-Leibler (KL) divergence​​, DKL(P∣∣Q)D_{KL}(P || Q)DKL​(P∣∣Q). It quantifies the "surprise" or information lost when you use a model distribution QQQ to approximate a true distribution PPP. It's not a perfect distance measure—crucially, DKL(P∣∣Q)≠DKL(Q∣∣P)D_{KL}(P || Q) \neq D_{KL}(Q || P)DKL​(P∣∣Q)=DKL​(Q∣∣P)—but for our purposes, it behaves like a squared distance.

Now for the grand finale. Imagine you have a complex "true" distribution PPP (the messy reality of your data) and a simpler family of models E\mathcal{E}E (say, the set of all possible bell curves). Finding the "best" approximation of PPP within E\mathcal{E}E is an act of projection: you are finding the point P∗P^*P∗ in the manifold E\mathcal{E}E that is "closest" to PPP in the KL sense. This is exactly what statisticians and machine learning algorithms do when they fit a model to data.

And, almost unbelievably, the Pythagorean theorem emerges once more. For the true distribution PPP, its best approximation P∗P^*P∗, and any other model QQQ in the family E\mathcal{E}E, a "Pythagorean theorem of information" holds:

DKL(P∣∣Q)=DKL(P∣∣P∗)+DKL(P∗∣∣Q)D_{KL}(P || Q) = D_{KL}(P || P^*) + D_{KL}(P^* || Q)DKL​(P∣∣Q)=DKL​(P∣∣P∗)+DKL​(P∗∣∣Q)

This is staggering. It means the total "error" in using an arbitrary model QQQ can be decomposed into two "orthogonal" parts: the "error" of the best possible model, DKL(P∣∣P∗)D_{KL}(P || P^*)DKL​(P∣∣P∗), and the "error" of moving from the best model to our arbitrary one, DKL(P∗∣∣Q)D_{KL}(P^* || Q)DKL​(P∗∣∣Q). This theorem is the conceptual backbone for many fundamental results in statistics and machine learning, guaranteeing that a certain kind of "orthogonality" holds when we find the best possible explanation for our data within a given class of models.

From a simple rule for triangles, the Pythagorean theorem has blossomed into a universal principle of orthogonality. It is a golden thread that ties together the geometry of space, the analysis of functions, the physics of fields, and the logic of inference. It reminds us that in mathematics, the simplest ideas are often the most profound, their echoes resounding in the most unexpected corners of knowledge.

Applications and Interdisciplinary Connections

We have explored the abstract skeleton of the generalized Pythagorean theorem, seeing how a simple idea about right triangles can be dressed in the elegant language of vectors and inner products. But what is the point of such abstraction? Does it do anything for us? I assure you, it does. This principle is not a museum piece to be admired from afar. It is a workhorse. It is a thread of geometric intuition that runs through nearly every branch of modern science and engineering. It appears in disguise, again and again, revealing a deep unity in the workings of the world. Let us now embark on a journey to spot this familiar ghost in some unexpected places.

The World in Many Dimensions

Our minds are comfortable with three spatial dimensions. But science and technology constantly force us to think in many more. A data scientist might describe a customer not by a position (x,y,z)(x, y, z)(x,y,z), but by a point in a 50-dimensional "feature space" whose axes are age, income, time spent on a website, items purchased, and so on. The "distance" between two customers in this space is a measure of their similarity, and it’s the bedrock of recommendation engines and targeted advertising. How do we measure this distance? With Pythagoras.

Imagine an nnn-dimensional hypercube, a perfect cube extended into nnn dimensions. How long is the grand diagonal that connects two opposite corners? If the side length is sss, we can imagine moving from one corner, (0,0,…,0)(0, 0, \dots, 0)(0,0,…,0), to the other, (s,s,…,s)(s, s, \dots, s)(s,s,…,s). This journey is equivalent to taking nnn consecutive steps, each of length sss, along nnn mutually perpendicular axes. The total displacement vector is (s,s,…,s)(s, s, \dots, s)(s,s,…,s). The Pythagorean theorem, generalized to nnn dimensions, tells us the squared length of this vector is simply the sum of the squares of its components: s2+s2+⋯+s2=ns2s^2 + s^2 + \dots + s^2 = n s^2s2+s2+⋯+s2=ns2. The distance is therefore sns\sqrt{n}sn​. An 11-dimensional cube with sides of 2.5 meters has a main diagonal of about 8.3 meters, a result computed with a tool forged over two millennia ago.

This principle of adding the squares of perpendicular components is universal. If we have a set of mutually orthogonal vectors, say v1,v2,…,vk\mathbf{v}_1, \mathbf{v}_2, \dots, \mathbf{v}_kv1​,v2​,…,vk​, the squared norm of their sum follows the same simple rule: ∥v1+v2+⋯+vk∥2=∥v1∥2+∥v2∥2+⋯+∥vk∥2\|\mathbf{v}_1 + \mathbf{v}_2 + \dots + \mathbf{v}_k\|^2 = \|\mathbf{v}_1\|^2 + \|\mathbf{v}_2\|^2 + \dots + \|\mathbf{v}_k\|^2∥v1​+v2​+⋯+vk​∥2=∥v1​∥2+∥v2​∥2+⋯+∥vk​∥2. This is the fundamental rule for combining independent quantities, and its echoes are everywhere.

The Art of Approximation and Decomposition

What happens when things are not perfectly aligned? In the real world, data is noisy and solutions are rarely perfect. Here, the Pythagorean theorem provides one of the most powerful concepts in all of quantitative science: orthogonal projection.

Imagine a flat plane, which we'll call the subspace WWW, inside our familiar 3D space. Think of this plane as representing all "possible" or "ideal" solutions to a problem. Now, suppose we have a vector a\mathbf{a}a that points somewhere outside this plane; this could be our messy, real-world data. We want to find the vector u\mathbf{u}u in the plane WWW that is "closest" to our data a\mathbf{a}a. The answer is to drop a perpendicular from the tip of a\mathbf{a}a down to the plane. The point where it lands is the tip of our best approximation, u\mathbf{u}u.

The vector connecting u\mathbf{u}u to a\mathbf{a}a, let's call it v=a−u\mathbf{v} = \mathbf{a} - \mathbf{u}v=a−u, is the "error" or "residual" vector. By its very construction, it is orthogonal to the plane WWW (and therefore to u\mathbf{u}u). We have decomposed our original data a\mathbf{a}a into two orthogonal parts: an ideal solution u\mathbf{u}u and an error v\mathbf{v}v. Since they are orthogonal, the Pythagorean theorem holds: ∥a∥2=∥u∥2+∥v∥2\|\mathbf{a}\|^2 = \|\mathbf{u}\|^2 + \|\mathbf{v}\|^2∥a∥2=∥u∥2+∥v∥2. The square of the "total length" is the sum of the square of the "solution length" and the square of the "error length". Minimizing the error ∥v∥\|\mathbf{v}\|∥v∥ is the whole idea behind the famous ​​method of least squares​​, which is used to fit lines to data points, analyze economic trends, and train countless machine learning models. The Pythagorean theorem is the geometric soul of statistical regression.

The Symphony of Signals and Waves

The journey of our theorem does not stop in finite dimensions. Let’s make a spectacular leap. What if our "vector" isn't a list of numbers, but a continuous entity, like a musical note held for one second, or the temperature distribution along a heated rod? These are functions, and Functions can be treated as vectors in an infinite-dimensional space called a Hilbert space.

The "inner product" of two functions is no longer a simple sum, but an integral. Two functions are "orthogonal" if the integral of their product over a given interval is zero. A beautiful example is the set of sine and cosine waves, {sin⁡(nx),cos⁡(nx)}\{\sin(nx), \cos(nx)\}{sin(nx),cos(nx)}, which are the building blocks of Fourier analysis. Waves of different integer frequencies are mutually orthogonal.

In this world, the Pythagorean theorem is reborn as ​​Parseval's Identity​​. It states that the total "energy" of a signal—defined as the integral of its squared value, ∫∣f(x)∣2dx\int |f(x)|^2 dx∫∣f(x)∣2dx—is equal to the sum of the energies of its individual orthogonal components. For a Fourier series, this means the total energy is the sum of the squares of the Fourier coefficients. This is a profound statement! It's the reason we can analyze a complex sound from a violin and talk meaningfully about the energy contained in its fundamental tone versus its overtones. This principle is the foundation of digital signal processing, enabling everything from audio compression in your music apps to image filtering in medical MRI scans. Furthermore, the geometric stability guaranteed by the Pythagorean structure is what ensures that infinite series of functions, like Fourier series, converge to a well-behaved limit, a cornerstone of mathematical analysis.

The Quantum Arena and the Logic of Information

The abstraction climbs higher still, and the rewards become even more profound. In the strange world of quantum mechanics, the state of a particle is described by a vector in a Hilbert space. Physical observables, like energy or momentum, are represented by special ​​Hermitian (or self-adjoint)​​ operators. The possible results of a measurement are the eigenvalues of these operators, and the system states corresponding to these definite outcomes are their eigenvectors, which form an orthonormal set.

When a particle is in a superposition of states, f=∑icivif = \sum_{i} c_i v_if=∑i​ci​vi​, what happens when we measure an observable TTT? The Pythagorean theorem's spirit guides the answer. The "average squared value" of the measurement, ⟨T2⟩\langle T^2 \rangle⟨T2⟩, is given by ∥Tf∥2\|Tf\|^2∥Tf∥2. Since the viv_ivi​ are orthonormal, the vectors Tvi=λiviTv_i = \lambda_i v_iTvi​=λi​vi​ are also orthogonal. Applying the theorem gives ∥Tf∥2=∑i∥λicivi∥2=∑i∣λi∣2∣ci∣2\|Tf\|^2 = \sum_i \|\lambda_i c_i v_i\|^2 = \sum_i |\lambda_i|^2 |c_i|^2∥Tf∥2=∑i​∥λi​ci​vi​∥2=∑i​∣λi​∣2∣ci​∣2. The squared "length" of the transformed state vector is the sum of the squared lengths of its components, weighted by squares of the measurement outcomes. The probabilities of obtaining each outcome λi\lambda_iλi​ are themselves given by ∣ci∣2|c_i|^2∣ci​∣2, a direct consequence of projecting the state vector onto the basis vectors. The entire probabilistic framework of quantum mechanics rests on this Hilbert space geometry.

Perhaps the most surprising appearance of our theorem is in the field of information theory. Here, the "distance" between two probability distributions ppp and qqq is often measured by a quantity called the Kullback-Leibler (KL) divergence, DKL(p∥q)D_{KL}(p\|q)DKL​(p∥q). It's not a true distance—it's not symmetric—but it behaves geometrically in a remarkably similar way. If you have a prior belief ppp and you receive new information that constrains your belief to a set C\mathcal{C}C, the optimal way to update your belief is to find the distribution q∗∈Cq^* \in \mathcal{C}q∗∈C that is "closest" to ppp. This q∗q^*q∗ is called an information projection.

A "generalized Pythagorean theorem" for information states that for any distribution rrr also in the constraint set C\mathcal{C}C, the "distance" from rrr to ppp decomposes perfectly: DKL(r∥p)=DKL(r∥q∗)+DKL(q∗∥p)D_{KL}(r\|p) = D_{KL}(r\|q^*) + D_{KL}(q^*\|p)DKL​(r∥p)=DKL​(r∥q∗)+DKL​(q∗∥p). This looks just like c2=a2+b2c^2 = a^2 + b^2c2=a2+b2! This is not just a mathematical party trick. This very property can be used to prove the convergence of complex, decentralized learning algorithms, where multiple agents must reach a consensus based on local information. The Pythagorean identity guarantees that the total "disagreement" in the system, measured by a sum of KL divergences, is a quantity that can only decrease with every step of communication, ensuring the system learns and stabilizes.

From the solid ground of geometry, to the noisy world of data, to the vibrating strings of a symphony, to the probabilistic haze of the quantum atom, and finally to the abstract logic of information itself—the Pythagorean theorem stands as a beacon. In its generalized form, it is far more than a formula. It is a fundamental principle of decomposition and harmony, dictating how to add up independent contributions, be they lengths, errors, energies, or even quantities of information. It is a stunning testament to the interconnectedness of all mathematics, and the power of a single, beautiful idea.