try ai
Popular Science
Edit
Share
Feedback
  • Sublinear Regret

Sublinear Regret

SciencePediaSciencePedia
Key Takeaways
  • Sublinear regret guarantees that an online algorithm's average performance converges to that of the best fixed action in hindsight.
  • Algorithms like Online Gradient Descent achieve O(√T) regret by taking small steps against the loss gradient with a carefully decaying step size.
  • Stronger problem assumptions, like strong convexity or data separability, allow for faster learning with logarithmic or even constant regret bounds.
  • The principle finds broad application in recommender systems, scientific discovery, game theory, and even provides a link to classical methods like the Kalman filter.

Introduction

In a world defined by uncertainty, how can we consistently make good decisions over time? Whether choosing which stock to buy, which route to take, or which scientific hypothesis to test, we are constantly forced to act with incomplete information. The goal cannot be to make the perfect choice every single day, as that would require knowing the future. This presents a fundamental challenge: what does it mean to 'learn' effectively in such an online setting, and how can we measure our progress?

This article addresses this question by exploring the powerful concept of ​​sublinear regret​​. Instead of aiming for impossible perfection, we adopt a more practical benchmark: comparing our cumulative performance to that of the best single fixed decision made in hindsight. Sublinear regret is the guarantee that, on average, the performance of our learning strategy will converge to that of this clairvoyant expert. It is a mathematical promise that we are bound to learn.

To understand this remarkable guarantee, we will embark on a two-part journey. The first chapter, ​​'Principles and Mechanisms'​​, will unpack the core algorithms, like Online Gradient Descent, that make sublinear regret achievable. We will investigate the crucial role of step sizes, explore how algorithms adapt to the problem's geometry, and discover how specific structural properties can lead to exponentially faster learning. Following this, the chapter ​​'Applications and Interdisciplinary Connections'​​ will reveal the astonishing reach of this principle, showcasing how it powers everything from personalized recommender systems and self-optimizing software to accelerating discovery in biotechnology and unifying disparate ideas in game theory and engineering.

Principles and Mechanisms

Imagine you are playing a game against the universe. Every day, you have to make a decision—which stock to invest in, which route to take to work, which drug candidate to synthesize. After you make your choice, the universe reveals how good it was by giving you a "loss". A high loss means a bad choice, a low loss means a good one. The catch is that the universe is unpredictable. The best stock today might be the worst tomorrow. You have no crystal ball. How can you possibly hope to play this game well?

This is the challenge of ​​online learning​​. Our goal isn't to be perfect on any given day. That's impossible. Instead, we play a more subtle game. We compare our total cumulative loss over many rounds, say TTT rounds, to that of a hypothetical, god-like expert. This expert is given a magical advantage: they get to see all TTT days in advance and pick the single best fixed decision to have made throughout the entire period. This fixed decision would have minimized their total loss in hindsight. The difference between our total loss and this expert's total loss is called the ​​regret​​.

Our goal is not to achieve zero regret; that would require time travel. Our humble, yet powerful, ambition is to ensure our regret grows slower than the number of rounds we play. We want our regret, RTR_TRT​, to be ​​sublinear​​ in TTT. This means that the average regret per round, RT/TR_T/TRT​/T, shrinks to zero as time goes on. In essence, we are guaranteed to learn. Over time, our performance, averaged out, will be just as good as that of the clairvoyant expert who knew the single best answer from the start. This is the remarkable promise of achieving sublinear regret. But how is such a feat possible?

The Simplest Strategy: A Step in the Right Direction

Let's think about the information we get each day. After we make our choice, say we pick a point wtw_twt​ from a set of possible decisions, the universe reveals a loss function, ℓt\ell_tℓt​. For now, let's assume this function is convex—shaped something like a bowl. This is a wonderful property, because it means that at our chosen point wtw_twt​, we can calculate a ​​gradient​​. The gradient is a vector that points "uphill," in the direction of the steepest increase in loss.

If we want to do better next time, the most natural thing to do is to take a small step in the direction opposite to the gradient. This is the core idea of ​​Online Gradient Descent (OGD)​​. At each round ttt, we update our decision for the next round, wt+1w_{t+1}wt+1​, by starting at our current decision wtw_twt​ and moving a little bit in the negative gradient direction:

wt+1′=wt−ηtgtw'_{t+1} = w_t - \eta_t g_twt+1′​=wt​−ηt​gt​

Here, gtg_tgt​ is the gradient of the loss ℓt\ell_tℓt​ at wtw_twt​, and ηt\eta_tηt​ is a small positive number called the ​​step size​​ or ​​learning rate​​, which controls how big a step we take. If our new point wt+1′w'_{t+1}wt+1′​ happens to fall outside our allowed set of decisions, we simply project it back to the nearest point within the set, giving us our final wt+1w_{t+1}wt+1​.

This seems almost too simple. Does it actually work? The magic lies in a beautiful piece of mathematical analysis. By tracking the distance between our iterate wtw_twt​ and the expert's optimal choice uuu, one can show that the regret after TTT rounds is bounded by a quantity that looks something like this:

RT(u)≤D22ηT+G22∑t=1TηtR_T(u) \le \frac{D^2}{2\eta_T} + \frac{G^2}{2}\sum_{t=1}^T \eta_tRT​(u)≤2ηT​D2​+2G2​t=1∑T​ηt​

Here, DDD is the "diameter" of our decision space (how far apart two possible decisions can be), and GGG is a bound on how steep the gradients can be.

Look closely at this formula. It reveals a fundamental tension, a trade-off at the heart of learning. The first term, D22ηT\frac{D^2}{2\eta_T}2ηT​D2​, gets smaller if our final step size ηT\eta_TηT​ is large. The second term, which involves a sum of all step sizes, gets smaller if our step sizes are small. To minimize our regret, we need to balance these two competing forces. The perfect compromise, it turns out, is to choose a step size that decays with time. A canonical choice is to set ηt\eta_tηt​ proportional to 1/t1/\sqrt{t}1/t​. With this choice, both terms in the bound end up growing like T\sqrt{T}T​. And so, we have it: the total regret RTR_TRT​ is on the order of T\sqrt{T}T​. This is sublinear! Our average regret, RT/T≈1/TR_T/T \approx 1/\sqrt{T}RT​/T≈1/T​, vanishes as TTT grows. This simple, intuitive strategy of following the negative gradient, with a carefully chosen decaying step size, is guaranteed to learn.

Adapting to the Terrain

The 1/t1/\sqrt{t}1/t​ step size is a universal solution, but it's a bit like wearing the same size shoes for every occasion. Sometimes the terrain is rougher in one direction than another. Imagine you are exploring a mountain range in the fog. At each step, you get a reading of the local steepness. If you consistently find that the north-south direction is extremely steep while the east-west direction is flat, you would naturally start taking smaller, more cautious steps when moving north or south, and larger, more confident steps when moving east or west.

Adaptive algorithms like ​​AdaGrad​​ do just this. Instead of using a pre-set decay schedule, they adapt the learning rate based on the history of gradients. The rule is simple: if the algorithm has seen large gradients in the past, it shrinks the learning rate for subsequent steps. A common way to do this is to set the learning rate at time ttt inversely proportional to the square root of the sum of squared norms of all gradients seen so far:

ηt∝1∑i=1t∥gi∥22+ϵ\eta_t \propto \frac{1}{\sqrt{\sum_{i=1}^t \|g_i\|_2^2 + \epsilon}}ηt​∝∑i=1t​∥gi​∥22​+ϵ​1​

where ϵ\epsilonϵ is a tiny number to prevent division by zero. This allows the algorithm to automatically "tune" itself to the geometry of the problem, often leading to much better practical performance than a fixed decay schedule.

When the World is Kinder: Finding Shortcuts to Wisdom

The O(T)O(\sqrt{T})O(T​) regret is a powerful guarantee because it holds even if the universe is being difficult (though not outright malicious). It assumes very little about the sequence of loss functions. But what if the world is more structured, or "kinder"? Can we learn even faster? The answer is a resounding yes.

The Separable World and the Perceptron

Consider a simple binary classification problem. At each round, we get a data point xtx_txt​ and a label yt∈{−1,+1}y_t \in \{-1, +1\}yt​∈{−1,+1}. Our job is to learn a weight vector www so that the sign of wTxtw^T x_twTxt​ matches the label yty_tyt​. What if the world is so kind that there exists a perfect "expert" vector uuu that correctly classifies every single data point with a comfortable margin of safety? That is, for every round ttt, yt(uTxt)≥γy_t(u^T x_t) \ge \gammayt​(uTxt​)≥γ for some positive margin γ>0\gamma > 0γ>0.

In this wonderfully structured world, we can use a remarkably simple algorithm: the ​​Perceptron​​. We start with w1=0w_1 = 0w1​=0. At each step, if we make a mistake, we update our weights by adding the misclassified point: wt+1=wt+ytxtw_{t+1} = w_t + y_t x_twt+1​=wt​+yt​xt​. If we get it right, we do nothing.

The analysis of this algorithm is one of the most elegant results in learning theory. By simultaneously tracking how the norm of our weight vector ∥wt∥2\|w_t\|^2∥wt​∥2 grows (it can only increase by so much at each mistake) and how aligned it becomes with the perfect expert uuu (it must become more aligned at each mistake), we can prove that the total number of mistakes the algorithm ever makes is finite and bounded by a constant:

Total Mistakes≤(Rγ)2\text{Total Mistakes} \le \left(\frac{R}{\gamma}\right)^2Total Mistakes≤(γR​)2

where RRR is the maximum size of a data point xtx_txt​. The total number of mistakes does not depend on the time horizon TTT at all! This means the regret is not just sublinear, it's O(1)O(1)O(1)—it becomes constant. After a finite number of updates, the algorithm achieves perfection and stops making mistakes. This beautiful result shows that with stronger assumptions comes the possibility of dramatically better performance.

The Curved World and Logarithmic Regret

Another way the world can be kind is if the loss functions have more "curvature". A standard convex function can have large flat regions, like the bottom of a wide, flat-bottomed canyon. In these regions, the gradient is small and provides little information about where the true minimum lies.

But what if the loss functions are ​​strongly convex​​? This means they are shaped like a nice, steep-sided bowl, with a clear, unique minimum. This extra curvature is a gift. It means that even far away from the minimum, the gradient provides a strong signal pointing toward it. Algorithms can exploit this structure to "zero in" on the target much more aggressively.

For such strongly convex problems, we can design algorithms that achieve a regret of O(log⁡T)O(\log T)O(logT). This is an exponential improvement over O(T)O(\sqrt{T})O(T​). For T=1,000,000T=1,000,000T=1,000,000, T\sqrt{T}T​ is 1,000, while log⁡T\log TlogT is only about 14. An excellent example is using a sophisticated method like the ​​Online Newton Step​​ for logistic regression, whose loss function has a property called exp-concavity, which is similar to strong convexity. This allows for logarithmic regret, whereas simpler methods on less-curved losses (like the hinge loss) are stuck with the T\sqrt{T}T​ rate.

Learning in the Dark: When Gradients are a Luxury

So far, we have assumed that after making our decision, we get to see the gradient. This tells us which direction we should have gone. But what if the feedback is more limited? What if we only learn the loss value at the single point we chose, and nothing else? This is called ​​bandit feedback​​. It's like a chef trying to perfect a recipe by only tasting the final cake, with no information about how changing the amount of sugar or flour would have affected the taste.

How can we possibly estimate a gradient with just one function value? The simplest idea is to poke the function in a random direction uuu and see what happens. We can form a ​​one-point estimator​​:

gt(1)=dδft(xt+δu)ug_t^{(1)} = \frac{d}{\delta} f_t(x_t+\delta u) ugt(1)​=δd​ft​(xt​+δu)u

Here, we perturb our point xtx_txt​ by a tiny amount δ\deltaδ in a random direction uuu and use the resulting function value to estimate the gradient. This estimator is unbiased (its average is the true gradient of a smoothed version of the function), but it is incredibly ​​noisy​​. Its variance, a measure of its noisiness, blows up like 1/δ21/\delta^21/δ2. This high variance pollutes our learning process, and the best regret we can hope for is a rather disappointing O(T3/4)O(T^{3/4})O(T3/4).

Can we do better? Yes, with a moment of cleverness. Instead of sampling one point, let's sample two points, symmetric around our current choice: xt+δux_t + \delta uxt​+δu and xt−δux_t - \delta uxt​−δu. We can then form a ​​two-point estimator​​ based on their difference:

gt(2)=d2δ(ft(xt+δu)−ft(xt−δu))ug_t^{(2)} = \frac{d}{2\delta}\big(f_t(x_t+\delta u)-f_t(x_t-\delta u)\big) ugt(2)​=2δd​(ft​(xt​+δu)−ft​(xt​−δu))u

This is a "central difference" approximation to the derivative, a familiar idea from calculus. The effect on variance is dramatic. Because we are taking a difference, the large, constant parts of the function value cancel out. The variance of this new estimator no longer depends on 1/δ21/\delta^21/δ2; in fact, it doesn't depend on δ\deltaδ at all! This much more stable gradient estimate allows us to recover the familiar O(T)O(\sqrt{T})O(T​) regret bound, even in the challenging bandit setting. It's a beautiful example of how a small change in algorithmic design can have a profound impact on performance.

For very expensive-to-evaluate functions, like in materials science or drug discovery, even two-point queries might be too much. Here, more sophisticated methods like ​​Bayesian Optimization​​ come into play. These algorithms build a full statistical model (like a Gaussian Process) of the unknown function, using both the value (mean) and the uncertainty (variance) to guide the search. The principle, known as "optimism in the face of uncertainty," is the same: balance exploiting what you know with exploring what you don't.

Learning with Style: The Quest for Simplicity

In many modern problems, especially in fields like genomics or finance, we deal with an enormous number of features. We might be trying to learn a model with millions of parameters. In such cases, we often prefer a simple or sparse solution—one where most of the parameters are exactly zero. A sparse model is easier to interpret, faster to evaluate, and less prone to overfitting.

Can we achieve sublinear regret while also encouraging sparsity? Yes, by using ​​composite losses​​. We add a regularization term to our loss, the most popular being the ℓ1\ell_1ℓ1​-norm, λ∥x∥1\lambda \|x\|_1λ∥x∥1​. The total loss becomes ft(x)+λ∥x∥1f_t(x) + \lambda \|x\|_1ft​(x)+λ∥x∥1​.

A standard gradient step would struggle with the kink in the ℓ1\ell_1ℓ1​-norm at zero. The solution is ​​Online Proximal Gradient Descent​​. The update is conceptually split in two. First, we take a normal gradient step on the smooth part of the loss, ftf_tft​. Then, we apply a special "proximal" step that handles the ℓ1\ell_1ℓ1​-norm. This proximal operator for the ℓ1\ell_1ℓ1​-norm is a simple and elegant function called ​​soft-thresholding​​:

(prox(v))i=sign(vi)max⁡{∣vi∣−θ,0}(\text{prox}(v))_i = \text{sign}(v_i) \max\{|v_i| - \theta, 0\}(prox(v))i​=sign(vi​)max{∣vi​∣−θ,0}

For each component of our vector, this operator subtracts a threshold θ\thetaθ from its magnitude, and if the result is negative, it sets the component to exactly zero. This is the key. It's an update rule that naturally produces sparse solutions. Remarkably, the regret analysis for this proximal method goes through just as before, yielding the standard O(T)O(\sqrt{T})O(T​) bound. We get the best of both worlds: a guarantee of learning and a tendency to find simple, sparse models.

A Final Word of Caution: What Regret Doesn't Tell You

We have seen that achieving sublinear regret is a powerful and general guarantee. It tells us that, on average, we will do as well as the best fixed decision in hindsight. But it is crucial to understand what this guarantee doesn't say.

Consider a multi-armed bandit problem where you must choose between two slot machines. One is slightly better than the other, but the difference in their average payout, the "gap" Δ\DeltaΔ, is very, very small. An algorithm like UCB (Upper Confidence Bound) is known to achieve sublinear regret. It will quickly learn to favor the better arm, but because the gap is so small, it will still feel the need to "check" the other arm occasionally, just to be sure. This exploration is necessary to keep the regret low.

Now, let's make the problem harder and harder by making the gap, ΔT\Delta_TΔT​, shrink as our time horizon TTT grows. We can construct a scenario where our regret is beautifully sublinear, perhaps growing like O(T1/4)O(T^{1/4})O(T1/4), yet the number of samples we would need to confidently identify which arm is better grows with time, perhaps like O(T1/2)O(T^{1/2})O(T1/2).

This reveals a subtle but critical distinction. Regret minimization is about ensuring good ​​cumulative performance​​ over a long period. Best-arm identification (a type of ​​PAC learning​​) is about finding the single best option with high confidence, as quickly as possible. Low regret does not automatically imply fast identification. An algorithm can be learning to perform well in an average sense, while still remaining uncertain about the absolute truth for a very long time, especially when the differences are faint. Understanding this distinction is key to wisely applying these powerful ideas in the real world.

Applications and Interdisciplinary Connections

"I can live with doubt and uncertainty and not knowing. I think it's much more interesting to live not knowing than to have answers which might be wrong." This sentiment, so beautifully expressed by Richard Feynman, is the very soul of scientific inquiry. But while we may live with uncertainty, we must still act. We must make decisions. The central question, then, is how to make the best possible decisions when we don't have all the facts.

In the last chapter, we uncovered a profound mathematical principle for doing just that: the idea of ​​sublinear regret​​. We found that it is possible to design strategies for repeated decision-making over a time horizon TTT that are, in a very precise sense, "guaranteed to learn." While we will inevitably make some mistakes along the way, the total cost of these mistakes—our cumulative regret—can be made to grow so slowly that our average per-round regret, 1T∑t=1T(lossour,t−lossoracle,t)\frac{1}{T} \sum_{t=1}^T (\text{loss}_{\text{our},t} - \text{loss}_{\text{oracle},t})T1​∑t=1T​(lossour,t​−lossoracle,t​), vanishes as time goes on. Our performance becomes indistinguishable from that of a mythical oracle who knew the right answer from the very beginning.

This is a powerful claim. But is it just a theoretical curiosity, confined to the abstract world of equations? Or is this principle at work in the world around us? We have forged a universal key. Now, let's take it for a walk and see just how many different doors it can unlock. The results, I think you will find, are quite astonishing.

The Digital World We Inhabit

Let’s start with something you likely experienced just a few minutes ago: the personalized internet. Every time you open a news app, a streaming service, or a social media feed, a decision is being made. Out of a universe of nnn articles, videos, or posts, which handful of kkk items should be presented to you? The service provider doesn't know your tastes perfectly. It faces a monumental "multi-armed bandit" problem. Each piece of content is a "slot machine," and your click is the "jackpot." How should it choose what to show you?

If it only ever shows you what it thinks you like based on past behavior (pure exploitation), it might miss a whole new genre you would have loved. If it only shows you random things to learn about you (pure exploration), your feed will feel chaotic and irrelevant. To be successful, the service must balance this trade-off. And the algorithms that do this best are precisely those that guarantee sublinear regret. They use principles like "optimism in the face of uncertainty" to explore intelligently, ensuring that the total regret—the cumulative difference between the engagement they got and the engagement they could have gotten with perfect knowledge of your taste—grows sublinearly. This means that over time, the system becomes an expert on you, and the cost of its initial ignorance becomes negligible. The smooth, personalized experience we now take for granted is, in many ways, a testament to the practical power of sublinear regret.

The Art of Computation Itself

Perhaps more surprisingly, this principle of learning is not just used to optimize our interaction with computers; computers can use it to optimize themselves. Think about a fundamental task, like multiplying two very large numbers of size nnn. There are different ways to do it. The method we learned in school is simple, but slow. The Karatsuba algorithm is faster, with a complexity of roughly O(nlog⁡23)O(n^{\log_2 3})O(nlog2​3). For truly enormous numbers, an algorithm based on the Fast Fourier Transform (FFT) is even faster, at around O(nlog⁡n)O(n \log n)O(nlogn).

Which one should a software library use? The answer depends on the size of the numbers. There are "crossover points" where one algorithm becomes better than another. We could try to figure these out by hand, but what if we frame it as a learning problem? We can treat each possible crossover point as an "expert" giving us advice. Every time the program needs to do a multiplication, it can, in a sense, ask the experts. By using a simple sublinear regret algorithm like the Multiplicative Weights Update method, the program can listen to these experts, see how well their advice performed, and dynamically update its trust in them. In a very short time, the system learns which expert to trust—that is, it finds the optimal crossover points automatically, based on its actual performance on the machine it's running on. Its total time spent is guaranteed to be nearly as good as if an oracle had told it the best settings from the start.

This idea can be pushed even further. The best algorithm for a task might not depend on something as simple as input size, but on a more subtle structural property of the data itself, like its Shannon entropy, H(P)H(P)H(P). By combining streaming algorithms that can estimate such properties "on the fly" with a learning framework, we can build adaptive programs that choose the best tool for the job in real time, again with provably minimal regret for making the wrong choice. The machine learns how to learn.

A New Engine for Scientific Discovery

The applications we've seen so far are remarkable, but they live inside a computer. What happens when the "actions" we take are not choosing bits, but manipulating atoms? The same principles apply, and the consequences are transformative for science.

Consider the challenge of directed evolution, a technique used to engineer new proteins for medicines or industrial enzymes. Scientists create vast libraries of mutated genes and then face the daunting task of finding the few "hits" with the desired properties. The experimental budget is always finite. Which mutants should you test? This is, again, a multi-armed bandit problem, but on a grand biological scale. Each variant library is an arm of the slot machine. An experimental test is a pull of the arm. A successful new protein is the jackpot. Sublinear regret algorithms like UCB (Upper Confidence Bound) and Thompson Sampling provide a formal recipe for allocating the precious experimental budget. They prescribe a precise way to balance testing the currently most promising variants (exploitation) with exploring the weird ones that have been neglected (exploration). This mathematical guidance dramatically accelerates the pace of discovery in biotechnology.

This paradigm extends to the most advanced frontiers of biology. Imagine you want to create a "minimal genome" for a bacterium by deleting all non-essential genes, a core challenge in synthetic biology. Or perhaps you're trying to find the perfect cocktail of dozens of growth factors to coax stem cells into forming a complex organoid, like a miniature brain in a dish. In both cases, the space of possibilities is astronomically large, and each experiment is incredibly expensive and time-consuming.

These are no longer simple bandit problems, as the choices are not independent. Deleting one gene might have a similar effect to deleting its neighbor on the chromosome. The effect of one growth factor depends smoothly on the concentration of another. Here, we can use more sophisticated learning models, like Gaussian Processes, that capture these correlations. But the core philosophy remains the same. The algorithm builds a probabilistic "surrogate" map of the unknown landscape. It then uses an acquisition function—a direct descendant of the UCB idea—to decide where to sample next, explicitly trading off exploiting known high-yield regions against exploring areas of high uncertainty. This family of techniques, known as Bayesian Optimization, is a direct application of the sublinear regret mindset. It allows scientists to find optimal genetic designs or differentiation protocols in dozens of experiments, where a brute-force approach would have required millions. It is a new engine for scientific discovery, turning uncertainty from an obstacle into a guide.

The Unseen Unity of Learning Systems

By now, you might be getting a sense of the universal nature of this idea. It seems to pop up everywhere. Let's conclude by looking at a few examples that reveal its role as a deep, unifying principle.

First, consider the complex, combinatorial choices we often face. It's not always about picking one best option, but about selecting a team of options to get a job done—like a logistics company choosing a daily set of delivery routes to cover all destinations, where the traffic on each route is unknown and fluctuating. This is a "combinatorial bandit" problem. Here too, the principle of "optimism in the face of uncertainty" can be extended to guide the company toward a near-optimal routing strategy, with total costs provably close to that of an all-knowing oracle.

Next, let's turn to game theory. Imagine two players in a competitive, zero-sum game. They don't know the other's strategy. How should they play? A fascinating result shows that if both players use a sublinear regret algorithm to update their strategy based on the outcomes of past rounds, the system as a whole will converge to a Nash Equilibrium—a stable state where neither player has an incentive to unilaterally change their strategy. The cumulative regret of each player is directly related to how far the game is from this equilibrium. It is a beautiful, decentralized mechanism for finding balance in a complex system. Even the geometry of the learning algorithm—whether it thinks in terms of Euclidean distances or information-theoretic divergences—subtly influences the path to equilibrium, a beautiful detail in a beautiful theory. This applies not just to board games, but to understanding dynamics in auctions, financial markets, and even ecological competition.

Finally, and perhaps most profoundly, consider the Kalman filter. For over half a century, this algorithm has been a cornerstone of engineering and applied science. It is the mathematical wizard behind the curtain in GPS receivers, weather forecasting models, and spacecraft navigation. It is a method for tracking a dynamically changing system (like a satellite's position) by continually blending a physical model's prediction with noisy measurements. It is the gold standard for data assimilation.

On the other hand, the field of online learning developed largely independently, focused on regret minimization for machine learning tasks. And yet, if you look under the hood, the two are deeply related. The core mathematical update of the Kalman filter can be shown to be identical to solving a particular online learning problem: an online version of ridge regression. And in the special case where the system being tracked is actually static, the Kalman filter's performance guarantees can be re-read in the language of sublinear regret. It turns out that this celebrated tool of engineering was, in a sense, a sublinear regret algorithm in disguise all along. What could be a more stunning example of the unity of scientific ideas? Two different communities, starting from different problems—tracking a missile versus predicting a user's click—arrived at the same fundamental mathematical structure.

From personalizing our news feeds, to optimizing our algorithms, to discovering new medicines, to finding balance in games, to tracking satellites in orbit—the principle of sublinear regret is a common thread. It is a formal theory of optimism and curiosity. It teaches us that while we cannot eliminate uncertainty, we can engage with it intelligently, ensuring that our journey through the unknown is an efficient one. The quest for sublinear regret is nothing less than the mathematical formalization of how to learn.