try ai
Popular Science
Edit
Share
Feedback
  • Infimum of Functions

Infimum of Functions

SciencePediaSciencePedia
Key Takeaways
  • The infimum of a set of functions is their greatest lower bound, constructed by taking the pointwise minimum value at every point in their shared domain.
  • The infimum operation preserves crucial mathematical properties, notably the measurability of functions and the geometric shape of concavity.
  • In optimization, the infimum describes the ultimate achievable limit, which may not be an attainable minimum, making it essential for analyzing ill-posed problems.
  • Fundamental quantities in science, such as ground-state energy in quantum physics and optimal strategies in control theory, are formally defined as an infimum over a vast space of possibilities.

Introduction

In mathematics and science, we are often concerned with finding the "best" or "lowest" value—be it the lowest energy state of a molecule or the minimum cost of a process. While the concept of a minimum is intuitive, reality often presents us with scenarios where a perfect minimum does not exist; instead, we can only approach an ultimate lower limit. This gap between an attainable minimum and an ideal lower bound is where the powerful concept of the ​​infimum of functions​​ becomes essential. This article demystifies the infimum, a fundamental tool for defining and understanding limits. The first chapter, "Principles and Mechanisms," will build the concept from the ground up, explaining how to construct an infimum and exploring its elegant mathematical properties. Following this, the "Applications and Interdisciplinary Connections" chapter will journey through various scientific fields, revealing how the infimum is used to define physical laws, solve complex optimization problems, and design optimal control systems. This journey will illuminate how a single abstract idea provides the language for some of science's most profound questions.

Principles and Mechanisms

Imagine you have a set of architectural blueprints for a roof, each designed by a different engineer. Some are steeply peaked, others are gently curved, and some are flat. Now, suppose you want to create a new design for a single, overarching structure that lies underneath all of these proposed roofs, hugging them as tightly as possible from below. This new surface, which represents the greatest possible lower boundary for all the designs, is the essence of the ​​infimum​​ of a collection of functions. It's a concept that is at once simple and profoundly powerful, acting as a foundational tool in fields from quantum mechanics to economic theory.

The Floor Beneath Your Feet: Defining the Infimum

Let's make this idea more concrete. In mathematics, we often define an order between functions. We say a function fff is "less than or equal to" a function ggg, written f⪯gf \preceq gf⪯g, if f(x)≤g(x)f(x) \le g(x)f(x)≤g(x) for every single point xxx in their domain. It's like saying one roof is entirely below or touching another at every point.

The ​​infimum​​ of a set of functions, which we can call hinfh_{\text{inf}}hinf​, is the "greatest" function that is less than or equal to every function in the set. It's the highest possible floor you can build underneath a collection of ceilings. How do we construct such a function? The secret is to think point by point. For any given xxx, the value of our infimum function, hinf(x)h_{\text{inf}}(x)hinf​(x), is simply the smallest value among all the functions in our set at that specific point xxx.

Consider a simple, hypothetical scenario: we have three different climate models, f1f_1f1​, f2f_2f2​, f3f_3f3​, predicting the temperature at three specific times of a day, say morning (x1x_1x1​), noon (x2x_2x2​), and evening (x3x_3x3​).

  • Model f1f_1f1​ predicts: {2,5,1}\{2, 5, 1\}{2,5,1}
  • Model f2f_2f2​ predicts: {4,0,3}\{4, 0, 3\}{4,0,3}
  • Model f3f_3f3​ predicts: {1,3,6}\{1, 3, 6\}{1,3,6}

To find the infimum function, hinfh_{\text{inf}}hinf​, which represents the absolute lowest temperature predicted by any model at each time, we just look at each time slot individually.

  • At morning (x1x_1x1​), the values are {2,4,1}\{2, 4, 1\}{2,4,1}. The minimum is 111. So, hinf(x1)=1h_{\text{inf}}(x_1) = 1hinf​(x1​)=1.
  • At noon (x2x_2x2​), the values are {5,0,3}\{5, 0, 3\}{5,0,3}. The minimum is 000. So, hinf(x2)=0h_{\text{inf}}(x_2) = 0hinf​(x2​)=0.
  • At evening (x3x_3x3​), the values are {1,3,6}\{1, 3, 6\}{1,3,6}. The minimum is 111. So, hinf(x3)=1h_{\text{inf}}(x_3) = 1hinf​(x3​)=1.

Our infimum function is therefore the set of predictions {1,0,1}\{1, 0, 1\}{1,0,1}. This ​​pointwise minimum​​ construction is the fundamental mechanism for finding the infimum. This same principle applies even in abstract settings, like Boolean logic. If you take two functions that are logical opposites—like p⊕qp \oplus qp⊕q (XOR, true when inputs differ) and p↔qp \leftrightarrow qp↔q (equivalence, true when inputs are same)—for any input, one is true (1) and the other is false (0). Their pointwise infimum is thus the constant function 000, a much simpler entity than either of its parents.

An Infinite Dance of Functions

The concept truly comes alive when we move from a small collection to an infinite sequence of functions. Imagine a sequence of functions, f1,f2,f3,…f_1, f_2, f_3, \dotsf1​,f2​,f3​,…, dancing on the real number line. The infimum function, g(x)=inf⁡nfn(x)g(x) = \inf_n f_n(x)g(x)=infn​fn​(x), traces the lower boundary of this infinite dance.

A beautiful and simple example is to consider a sequence of "shrinking boxes." Let fn(x)f_n(x)fn​(x) be a function that is 111 inside the open interval (−1n,1n)(-\frac{1}{n}, \frac{1}{n})(−n1​,n1​) and 000 everywhere else. This is known as a characteristic function, χ(−1/n,1/n)(x)\chi_{(-1/n, 1/n)}(x)χ(−1/n,1/n)​(x).

  • f1(x)f_1(x)f1​(x) is 111 on (−1,1)(-1, 1)(−1,1).
  • f2(x)f_2(x)f2​(x) is 111 on (−12,12)(-\frac{1}{2}, \frac{1}{2})(−21​,21​).
  • f100(x)f_{100}(x)f100​(x) is 111 on (−1100,1100)(-\frac{1}{100}, \frac{1}{100})(−1001​,1001​).

What is the infimum of this sequence? Let's pick a point xxx and ask. If x=0.1x = 0.1x=0.1, then for n=1,2,…,9n=1, 2, \dots, 9n=1,2,…,9, fn(0.1)=1f_n(0.1) = 1fn​(0.1)=1. But for n=10n=10n=10 and all larger integers, 0.10.10.1 is no longer in the interval (−1n,1n)(-\frac{1}{n}, \frac{1}{n})(−n1​,n1​), so fn(0.1)=0f_n(0.1)=0fn​(0.1)=0. Since the sequence of values {1,1,…,1,0,0,0,… }\{1, 1, \dots, 1, 0, 0, 0, \dots\}{1,1,…,1,0,0,0,…} contains a zero, its infimum (greatest lower bound) is 000. This is true for any x≠0x \neq 0x=0. No matter how small xxx is, we can always find a large enough nnn such that 1n∣x∣\frac{1}{n} |x|n1​∣x∣, pushing the value of fn(x)f_n(x)fn​(x) to zero.

The only exception is the point x=0x=0x=0. The number 000 is inside every interval (−1n,1n)(-\frac{1}{n}, \frac{1}{n})(−n1​,n1​). Thus, for x=0x=0x=0, the sequence of values is {1,1,1,… }\{1, 1, 1, \dots\}{1,1,1,…}. The infimum of this sequence is 111. So, the infimum function is a strange beast: it is 111 at the single point x=0x=0x=0 and 000 everywhere else. It's the characteristic function of the set containing only zero, χ{0}(x)\chi_{\{0\}}(x)χ{0}​(x). The infimum has captured the limiting behavior of the sequence of intervals, which shrink to a single point.

Sometimes the dance is more complex. Consider the sequence fn(x)=(1+(−1)nn)x2f_n(x) = \left(1 + \frac{(-1)^n}{n}\right)x^2fn​(x)=(1+n(−1)n​)x2. The coefficient cn=1+(−1)nnc_n = 1 + \frac{(-1)^n}{n}cn​=1+n(−1)n​ oscillates as it converges to 1. For even nnn, it approaches 111 from above (e.g., 32,54,…\frac{3}{2}, \frac{5}{4}, \dots23​,45​,…), while for odd nnn, it approaches 111 from below (e.g., 0,23,45,…0, \frac{2}{3}, \frac{4}{5}, \dots0,32​,54​,…). To find the infimum of the sequence fn(x)f_n(x)fn​(x), we just need the infimum of the coefficients cnc_ncn​, since x2≥0x^2 \ge 0x2≥0. The lowest value ever reached by cnc_ncn​ is at n=1n=1n=1, where c1=0c_1=0c1​=0. Therefore, the infimum function is h(x)=0⋅x2=0h(x) = 0 \cdot x^2 = 0h(x)=0⋅x2=0. The upper boundary, the supremum, would be set by the largest coefficient, which is c2=32c_2 = \frac{3}{2}c2​=23​, giving a supremum function g(x)=32x2g(x) = \frac{3}{2}x^2g(x)=23​x2.

A Secret of Numbers Revealed

The true magic of the infimum appears when we probe the very fabric of the number line. Let's look at the sequence fn(x)=cos⁡2(πnx)f_n(x) = \cos^2(\pi n x)fn​(x)=cos2(πnx) on the interval [0,1][0, 1][0,1]. Since cos⁡2\cos^2cos2 can never be negative, the infimum must be greater than or equal to zero. The question is: can we always make it zero?

The answer depends, astonishingly, on whether xxx is rational or irrational.

  • ​​Rational Case 1:​​ Take x=310x = \frac{3}{10}x=103​. Can we make fn(310)=cos⁡2(3πn10)f_n(\frac{3}{10}) = \cos^2(\frac{3\pi n}{10})fn​(103​)=cos2(103πn​) equal to zero? Yes! We need the argument inside the cosine to be an odd multiple of π2\frac{\pi}{2}2π​, like π2,3π2,…\frac{\pi}{2}, \frac{3\pi}{2}, \dots2π​,23π​,…. Choosing n=5n=5n=5, the argument becomes 15π10=3π2\frac{15\pi}{10} = \frac{3\pi}{2}1015π​=23π​. And cos⁡(3π2)=0\cos(\frac{3\pi}{2})=0cos(23π​)=0. Since one of the functions in the sequence hits zero, the infimum is 000.

  • ​​Rational Case 2:​​ Now take x=35x = \frac{3}{5}x=53​. We are looking for an integer nnn that makes 3πn5\frac{3\pi n}{5}53πn​ an odd multiple of π2\frac{\pi}{2}2π​. This requires 6n=5×(an odd number)6n = 5 \times (\text{an odd number})6n=5×(an odd number), which is impossible since the left side is even and the right side is odd. The function value never reaches zero! In this case, the sequence of values fn(3/5)f_n(3/5)fn​(3/5) is periodic and cycles through a small, finite set of positive numbers. The infimum is simply the smallest of these, which turns out to be cos⁡2(2π5)=3−58\cos^2(\frac{2\pi}{5}) = \frac{3 - \sqrt{5}}{8}cos2(52π​)=83−5​​.

  • ​​Irrational Case:​​ What if xxx is irrational, say x=2−1x = \sqrt{2}-1x=2​−1? Now we can never find an integer nnn to make πn(2−1)\pi n (\sqrt{2}-1)πn(2​−1) an exact odd multiple of π2\frac{\pi}{2}2π​, because that would imply 2\sqrt{2}2​ is rational. So does that mean the infimum is positive? No! Here we witness a deep property of numbers. A famous result in mathematics (related to Weyl's criterion) states that the sequence of fractional parts of the multiples of an irrational number, {nx(mod1)}\{nx \pmod 1\}{nx(mod1)}, is dense in the interval [0,1][0, 1][0,1]. This means you can get arbitrarily close to any number in that interval, including 12\frac{1}{2}21​. This allows us to find a sequence of integers nkn_knk​ such that nkxn_k xnk​x gets closer and closer to a half-integer (like 0.5,1.5,…0.5, 1.5, \dots0.5,1.5,…). Consequently, cos⁡(πnkx)\cos(\pi n_k x)cos(πnk​x) gets arbitrarily close to 000. While no function in the sequence may ever be exactly zero, they get so close that their greatest lower bound—the infimum—is precisely 000.

So, the infimum function g(x)=inf⁡ncos⁡2(πnx)g(x) = \inf_n \cos^2(\pi n x)g(x)=infn​cos2(πnx) acts as a detector for number types: it is 000 for almost all inputs (all irrationals and some rationals) but pops up to positive values on a fine dust of specific rational numbers.

The Rules of the Game: Why the Infimum is a Reliable Tool

Beyond these fascinating behaviors, mathematicians and physicists prize the infimum because it's a well-behaved and reliable operation. In modern physics and analysis, we often work with "measurable" functions. Informally, a function being ​​measurable​​ means you can ask a question like, "Where is this function's value greater than aaa?" and the resulting set of points is a "sensible" region (a measurable set) for which you can define concepts like area or volume.

A critical property is that if you start with a sequence of measurable functions, their infimum is also measurable. The proof is elegant and reveals the infimum's connection to fundamental set operations.

  • For the infimum g(x)=inf⁡nfn(x)g(x) = \inf_n f_n(x)g(x)=infn​fn​(x) to be ​​greater than or equal to​​ a value aaa, it must be that every single function fn(x)f_n(x)fn​(x) is also greater than or equal to aaa. A chain is only as strong as its weakest link; if even one fk(x)f_k(x)fk​(x) dips below aaa, the infimum will be at most fk(x)f_k(x)fk​(x), so it cannot be ≥a\ge a≥a. This means the set where g(x)≥ag(x) \ge ag(x)≥a is the ​​intersection​​ of all the sets where fn(x)≥af_n(x) \ge afn​(x)≥a.
  • For the infimum g(x)g(x)g(x) to be ​​less than​​ a value aaa, we only need at least one function fn(x)f_n(x)fn​(x) to dip below aaa. If you can find just one plank in the floor that is below height aaa, then the "greatest lower bound" must also be below aaa. This means the set where g(x)ag(x) ag(x)a is the ​​union​​ of all the sets where fn(x)af_n(x) afn​(x)a.

Since measurable sets are closed under countable unions and intersections (this is part of the definition of a σ\sigmaσ-algebra, the collection of all "sensible" sets), these relationships guarantee that the infimum of measurable functions is measurable. This closure property is essential—it allows us to build complex functions from simple ones without leaving the well-behaved world of measurable functions.

Furthermore, the infimum is robust when dealing with the concept of "almost everywhere" equality. In measure theory, we often consider two functions to be equivalent if they only differ on a set of measure zero—a set of "dust" like the rational numbers on the real line. If you take two sequences of functions {fn}\{f_n\}{fn​} and {gn}\{g_n\}{gn​} that are equal almost everywhere for each nnn, their respective infima, fff and ggg, will also be equal almost everywhere. The infimum operation respects this equivalence and isn't thrown off by negligible differences.

Shaping a New Landscape: Infima, Convexity, and Concavity

Finally, what does the infimum do to the shape of functions? If we take the infimum of functions with a nice geometric property, does the result inherit that property?

Consider ​​convex functions​​—functions that are "bowl-shaped," like f(x)=x2f(x)=x^2f(x)=x2. The line segment connecting any two points on their graph always lies above or on the graph. Does the infimum of two convex functions have to be convex? The answer is a resounding no. Imagine two bowls, f(x)=(x−2)2f(x)=(x-2)^2f(x)=(x−2)2 and h(x)=(x+2)2h(x)=(x+2)^2h(x)=(x+2)2. One is centered at x=2x=2x=2, the other at x=−2x=-2x=−2. Their infimum follows the curve of h(x)h(x)h(x) for x0x0x0 and switches to f(x)f(x)f(x) for x>0x>0x>0. The resulting shape looks like a "W," with a central peak at x=0x=0x=0. This "W" shape is decidedly not a single bowl—it's not convex.

But here comes a moment of beautiful symmetry. What about ​​concave functions​​—functions that are "mound-shaped," like f(x)=−x2f(x) = -x^2f(x)=−x2? If you take the pointwise infimum of any family of concave functions, the resulting function is ​​always concave​​. You can picture this: if you have a collection of mounds and you trace their greatest lower boundary, you get another, possibly more complex, mound-like shape with sharp ridges where it switches from following one function to another. This preservation of concavity is a cornerstone of optimization theory and economics, where finding the maximum of a concave function (a desirable task) is guaranteed to yield a global, not just local, optimum.

From a simple pointwise rule to a subtle probe of the number line and a tool that sculpts new functions with predictable geometric properties, the infimum is a concept of remarkable depth and utility. It is one of the fundamental operations that allows us to build the intricate and powerful structure of modern mathematical analysis.

Applications and Interdisciplinary Connections

Now that we have a feel for the delicate dance between a minimum and an infimum, you might be wondering, "Is this just a clever distinction for mathematicians to ponder?" Not at all! This idea is not some dusty relic in a cabinet of curiosities. It is a sharp, powerful tool that scientists and engineers use to ask some of the most profound questions about the world. It is the language we use to talk about ultimate limits, fundamental states, and optimal strategies. Let's go on a little tour and see where this concept lives and breathes.

The Quest for the "Best": Optimization and the Real World

At its heart, much of science and engineering is a grand optimization problem. We want to build the strongest bridge with the least material, design a drug with the maximum effect and minimum side effects, or run a factory at the lowest cost. In the perfect world of a textbook, every problem has a perfect answer—a "minimum." But the real world is often not so tidy.

Imagine you are an engineer tuning two different systems. In one, you're balancing the cost of running a machine for a long time against the cost of rushing the process. The total cost might look something like C1(t)=At+BtC_1(t) = \frac{A}{t} + BtC1​(t)=tA​+Bt. If you run it too fast (small ttt), the first term blows up; too slow (large ttt), and the second term dominates. Common sense, and calculus, tells us there must be a sweet spot, a perfect time t∗t^*t∗ that gives the absolute minimum cost. This problem is "well-posed"; an optimal solution exists and can be found. The infimum is a minimum.

But now consider a second system, where you are trying to minimize the concentration of a decaying catalyst, which follows a rule like C2(t)=Dexp⁡(−kt)C_2(t) = D \exp(-kt)C2​(t)=Dexp(−kt). The process starts at t=0t=0t=0, and you want to find the minimum concentration after it begins, i.e., for t>0t > 0t>0. The concentration is always decreasing, getting closer and closer to zero as time goes on. What is the minimum concentration? Well, you can get it as low as you want by waiting long enough, but you can never actually reach a concentration of zero in any finite time. The greatest lower bound—the infimum—is 0, but this value is never attained. There is no "best" time to stop; there's only "better." This problem is "ill-posed" for a minimum, and the infimum is the only tool that can precisely describe this asymptotic goal. This distinction is not academic; it tells an engineer whether they are searching for a specific setting or chasing an unreachable ideal.

The existence of a minimum often depends critically on the world, or domain, you are allowed to search in. Suppose you have a function and are looking for its lowest point. If your searching ground is a closed, bounded region—a "compact" set, in mathematical terms—and your function is continuous, the Extreme Value Theorem guarantees you will find a minimum. It’s like searching for the lowest point in a fenced-off, finite valley; there's definitely a bottom.

But what if your domain is more peculiar? What if you are only allowed to stand on discrete lily pads (like the rational numbers) scattered across the valley? You might see the true bottom of the valley between two lily pads, a point corresponding to an irrational number. You can hop from one pad to the next, getting your feet arbitrarily close to the bottom, but you can never stand right on it. The infimum of your altitude would be the true bottom of the valley, but you would never attain it as a minimum. The infimum tells you the limit of what's possible, even when the rules of the game prevent you from getting there.

Defining the Fabric of Reality: From Physical Laws to Abstract Spaces

Perhaps the most beautiful use of the infimum is not in finding a value, but in defining one. Many of the fundamental constants and quantities in nature are, in fact, the answer to an infinite-dimensional optimization problem, expressed as an infimum.

Think of the sound a drum makes. A drumhead of a certain shape, when struck, can vibrate in many different ways, or modes, each with its own frequency. But there is a lowest possible frequency, its fundamental tone. How do we find this tone? Mathematical physics tells us that this fundamental frequency corresponds to a quantity called the first Dirichlet eigenvalue, λ1\lambda_1λ1​. And how is λ1\lambda_1λ1​ defined? It is the infimum of the "Rayleigh quotient"—a ratio of the drumhead's bending energy to its displacement—taken over all possible smooth shapes the drumhead could form. Nature, in its essence, is lazy. When it vibrates, it seeks the path of least resistance, the mode with the lowest energy-to-displacement ratio. The fundamental frequency is this infimum. It isn't calculated from a simple formula; it is the ultimate lower bound for an infinite family of possibilities.

This principle echoes through the deepest level of physics: quantum mechanics. The holy grail for a chemist is to find the "ground-state energy" of a molecule—its lowest possible energy, which determines its stability, shape, and reactivity. The modern way to do this, called Density Functional Theory (DFT), is a masterpiece of the infimum concept. The ground-state energy is found by minimizing an energy functional over all possible electron densities. But what is the functional itself? A key piece of it, the universal functional F[ρ]F[\rho]F[ρ], is defined through a "constrained search": it is the infimum of the kinetic and interaction energy over the set of all possible quantum wavefunctions that could give rise to a specific electron density ρ\rhoρ. It’s a mind-bendingly abstract question: "If the electrons were arranged to create this cloud-like density ρ\rhoρ, what's the absolute minimum internal energy they could have?" The answer defines the functional. This framework even has a clever, built-in reality check. What if we propose a "density" that is physically impossible, say, one that is negative in some region? No wavefunction can create it. The set of wavefunctions to search over is empty. By convention, the infimum over an empty set is +∞+\infty+∞. This elegantly ensures that the theory automatically rejects unphysical nonsense and only considers real possibilities.

The power of the infimum extends even to characterizing the very nature of abstract mathematical spaces. In functional analysis, which studies infinite-dimensional spaces, one can ask about the "shape" of the unit ball. Is it perfectly round like a soccer ball, or does it have flat spots like a cut diamond? A quantity called the "modulus of convexity" measures this. It is defined as an infimum that captures how much the midpoint between any two points on the surface of the ball "sags" toward the center. An infimum of zero suggests a flat spot, while a larger value implies a nice, uniform roundness. Here, the infimum isn't just finding a single number; it's defining a geometric character trait of an entire universe of functions.

The Dynamics of Control: Steering Toward an Optimum

Finally, let’s look at the world of control theory, where things change in time. Whether you're designing a self-driving car, a robot arm, or a thermostat, the goal is to make decisions over time to achieve an objective. The principle of optimality, which underpins this entire field, is written in the language of the infimum.

Imagine you want to stabilize a system described by x˙=f(x)+g(x)u\dot{x} = f(x) + g(x)ux˙=f(x)+g(x)u, where uuu is your control input (the steering, the throttle). You have a function V(x)V(x)V(x) that measures how "bad" the current state xxx is; you want to drive V(x)V(x)V(x) to zero. At any instant, the rate of change of VVV depends on your choice of uuu. To do the best possible job, you should choose the control uuu that makes V˙\dot{V}V˙ decrease as quickly as possible. That is, you want to find the infimum of V˙\dot{V}V˙ with respect to uuu. This very calculation is at the heart of designing a Control Lyapunov Function (CLF). The analysis reveals something beautiful: if the term multiplying your control, LgV(x)L_gV(x)Lg​V(x), is non-zero, you have authority and can, in principle, make V˙\dot{V}V˙ arbitrarily negative (the infimum is −∞-\infty−∞). If it happens to be zero, you've hit a point where your controls have no instantaneous effect on VVV, and the best you can do is accept the natural drift of the system.

This idea reaches its zenith in the Hamilton-Jacobi-Bellman (HJB) equation, the master equation of optimal control. The entire equation is built around a "Hamiltonian," which is defined as an infimum of the costs and system dynamics over all possible control actions one could take at a given moment. Solving the HJB equation is like knowing the "optimal value" of being in any state at any time, assuming you will act optimally from that point forward. That future optimal action is encoded in the infimum. And, just as we saw before, questions about whether an optimal control exists at every moment often come down to the properties of the control set. If the set of available controls is compact (closed and bounded), the Weierstrass theorem ensures that a "best" decision can always be made.

From the factory floor to the shape of abstract universes, from the tone of a drum to the ground state of a molecule, the concept of the infimum is there. It is the language of limits, of bounds, of the fundamental and the optimal. It is a testament to the power of a single, precise idea to unify our understanding of a wonderfully complex world.