try ai
Popular Science
Edit
Share
Feedback
  • The Power of the Outlier: An Introduction to Heavy-Tailed Models

The Power of the Outlier: An Introduction to Heavy-Tailed Models

SciencePediaSciencePedia
Key Takeaways
  • Heavy-tailed distributions decay via power laws, making extreme outliers far more probable and impactful than in thin-tailed distributions like the Normal curve.
  • In systems governed by heavy tails, traditional statistics like the mean and standard deviation are often misleading or infinite, making robust metrics like the median essential.
  • Heavy tails arise from specific physical mechanisms, like multiplicative processes or mixtures of rates, and are a unifying feature in phenomena from biological evolution to computer network latency.
  • Concepts from Extreme Value Theory, rather than the Central Limit Theorem, provide the correct mathematical framework for understanding and predicting the maximums in heavy-tailed data.

Introduction

Our world is often summarized by averages, governed by the familiar bell curve where extreme events are vanishingly rare. This intuition, however, breaks down in many critical systems—from financial markets to biological evolution—where a single outlier can dominate the entire picture. These are the realms of heavy-tailed distributions, a concept that challenges our reliance on traditional statistics and forces us to confront the profound impact of the extreme. This article provides an essential guide to this counter-intuitive world. It addresses the knowledge gap between our well-behaved statistical training and the wild reality of many natural and engineered systems.

In the first chapter, "Principles and Mechanisms," we will explore the fundamental properties of heavy tails, understand why concepts like the average fail, and discover the physical processes that give rise to them. Following this, the "Applications and Interdisciplinary Connections" chapter will take us on a journey across diverse scientific fields, revealing how the signature of the heavy tail provides a unifying language to understand phenomena from brain structure to computer performance. By the end, you will have a new lens through which to view data, one that is prepared for the power of the outlier.

Principles and Mechanisms

Imagine you are mapping a newly discovered country. You might start by finding its center, its most populous city, and then measure how the population thins out as you travel toward the borders. For most countries we know, the population density drops off rather quickly. A few miles out of the city, you're in the suburbs; a few more, and you're in the countryside; and soon after, you find yourself in the vast, empty wilderness. The borders are reached relatively quickly. Distributions like this, such as the famous bell curve or Normal distribution, are "thin-tailed." The probability of finding someone extremely far from the center dwindles incredibly fast—exponentially fast, in fact.

But what if you discovered a country of a completely different sort? A country where, long after you've left the capital, you keep stumbling upon significant, thriving towns in the distant hinterlands. The population density decreases, but far more slowly, following a more gradual slope. This is the world of "heavy-tailed" distributions. They are the lands of the unexpected, the domains where black swans are native birds. The "tail" of the distribution, representing the probability of extreme events, is far heavier, or fatter, than we might intuitively expect. It doesn't fall off a cliff; it stretches on, long and massive, like the tail of a mythological dragon.

The Tale of the Dragon's Tail: What is a "Heavy Tail"?

Let’s be a little more precise. The difference between a thin and a heavy tail is the rate at which the probability of extreme events goes to zero. A thin-tailed distribution, like the Gaussian, decays exponentially. The probability of an event far from the mean, say at a distance xxx, might be proportional to exp⁡(−x2)\exp(-x^2)exp(−x2). This number shrinks with astonishing speed. An event that is 10 standard deviations from the mean is already fantastically improbable.

A heavy-tailed distribution, on the other hand, decays much more leisurely, typically following a power law. Here, the probability of an event larger than xxx might be proportional to x−αx^{-\alpha}x−α for some positive exponent α\alphaα. While this also goes to zero as xxx gets larger, it does so far more slowly than any exponential. This seemingly subtle mathematical difference has earth-shattering consequences.

Nature, it turns out, is full of such heavy tails. Consider the delicate dance of quantum mechanics within an atom. In quantum chemistry, we often try to approximate the true wavefunction of an electron—which describes where the electron is likely to be—using a combination of simpler Gaussian functions, exp⁡(−αr2)\exp(-\alpha r^2)exp(−αr2). However, the actual wavefunction of a loosely bound electron, like one in a negative ion, decays more slowly, something like exp⁡(−κr)\exp(-\kappa r)exp(−κr). Compared to the breakneck speed at which a Gaussian function plunges to zero, this simple exponential decay is a gentle giant. It is, in effect, "heavy-tailed" relative to our Gaussian building blocks. To accurately capture this slowly decaying tail, which contains a significant amount of the electron's probability, chemists must add special "diffuse" functions to their models—Gaussian functions with very small exponents α\alphaα that are themselves spread out and long-ranged. Without them, properties that depend on the electron's periphery, like the ability to capture another electron or the response to an electric field, are calculated incorrectly. The heavy tail is not a mathematical curiosity; it is a physical reality.

The Failure of the Average and the Rise of the Robust

For centuries, our statistical thinking has been dominated by the thin-tailed world. We revere the ​​mean​​, or average, as the ultimate summary of a set of data. We use the standard deviation to tell us about the typical spread. This toolkit works wonderfully when data comes from a distribution like the Normal curve. The Central Limit Theorem—a titan of probability theory—tells us that if you add up enough independent random things, their sum will tend to follow a Normal distribution, regardless of what the individual things looked like. This theorem gives us a sense of safety, a belief in the eventual triumph of the average.

But in the land of heavy tails, this safety is an illusion. The Central Limit Theorem has a crucial requirement in its fine print: the things you are adding up must have a finite variance (a finite standard deviation). Many heavy-tailed distributions, particularly those with a power-law exponent α≤2\alpha \le 2α≤2, violate this condition. Their variance is literally ​​infinite​​.

What does this mean? It means there is no "typical" spread. The outliers are so extreme, and occur just often enough, that the measure of average squared deviation from the mean never settles down. As you collect more data, a new, even wilder outlier is always lurking, ready to come along and blow up your calculation of the variance.

Imagine you are a computer architect measuring the performance of a processor by running a benchmark multiple times. You measure the "cycles per instruction" (CPI). Most of the time, the value is around 1.011.011.01 or 1.021.021.02. But the machine is shared, and occasionally, the operating system or another program interferes, causing a massive delay. In a set of twelve runs, you might see values like: 1.00,1.01,1.02,…,2.60,3.901.00, 1.01, 1.02, \dots, 2.60, 3.901.00,1.01,1.02,…,2.60,3.90. If you calculate the average CPI, the two huge outliers will drag the result up to around 1.381.381.38. Does this number represent the "typical" performance? Not at all. It's a distorted picture, tyrannized by the outliers. This is a classic symptom of a system with heavy-tailed performance jitter.

In this world, we need a new hero. That hero is the ​​median​​. The median is simply the middle value of the sorted data. For our CPI data, the median is a sensible 1.0151.0151.015. It captures the central tendency of the bulk of the data, because it is fundamentally immune to the magnitude of outliers. You could change the 3.903.903.90 to a million, and the median would not budge. Statisticians call this property ​​robustness​​. The median has a high "breakdown point"—you can corrupt almost 50% of your data without sending the median to an absurd value. The mean, by contrast, has a breakdown point of zero; a single bad data point can destroy it.

This same principle applies when we evaluate the performance of predictive models. In materials science, if we are predicting a material's properties, our model will have errors. If the error distribution is heavy-tailed, some predictions will be spectacularly wrong. If we measure our model's overall performance using the Root Mean Squared Error (RMSE), which involves squaring the errors, these few huge errors will dominate the entire metric and give a pessimistic and unstable assessment. A much more robust metric is the Mean Absolute Error (MAE), which is not squared. The MAE behaves more like the median, giving a more stable and often more useful picture of the model's typical accuracy in the face of these outliers.

The Birth of Giants: How Nature Forges Heavy Tails

If heavy tails are so different, where do they come from? They are not just arbitrary mathematical functions; they often arise from specific, understandable physical mechanisms.

Let's return to a molecular machine: RNA polymerase, the enzyme that reads our DNA to create RNA. We can watch a single molecule as it chugs along the DNA template. Most of the time, it moves along at a steady pace. But sometimes, it pauses. If we measure the duration of these pauses, we find that the distribution has a long, power-law tail. Why?

One simple model of a process is a sequence of steps: step 1, then step 2, ..., then step NNN. If each step is a random, memoryless waiting process (an exponential distribution), the total time will be the sum of these waiting times. But as we've seen, adding up such random variables, by the logic of the Central Limit Theorem, tends to create a distribution that is decidedly thin-tailed. A simple sequential process cannot produce a heavy tail.

Nature must be more clever. The polymerase, it turns out, doesn't just pause; it can enter a completely different "off-pathway" state. For instance, it might backtrack along the DNA. To resume its work, it must diffuse back to the correct position. The time it takes to return by a random, one-dimensional walk—a process physicists call a "first-passage time"—is not exponentially distributed. Its distribution has a heavy, power-law tail decaying as t−3/2t^{-3/2}t−3/2. A single diffusive mechanism can give birth to a heavy-tailed waiting time.

There is another, equally profound way. Imagine the polymerase can enter many different kinds of paused states. Some are shallow and easy to escape, with a high escape rate kkk. Others are deep and stable, with a very low escape rate. If the observed pause time is a random choice from a whole menu of different exponential processes, the resulting mixture is no longer a simple exponential. The overall survival probability S(t)S(t)S(t) is an average of many different exp⁡(−kt)\exp(-kt)exp(−kt) terms. The ones with large kkk die out quickly. But the rare, slow-to-escape states, those with kkk close to zero, linger. Their contribution, exp⁡(−kt)≈1−kt\exp(-kt) \approx 1 - ktexp(−kt)≈1−kt, dominates at long times and, when averaged, can beautifully stitch together a perfect power-law tail. This idea—that a ​​mixture of simple exponential processes can create a complex power law​​—is a deep and recurring theme in physics and biology.

Accelerating Waves and the Law of the Largest

The influence of heavy tails extends beyond static measurements to dynamic processes, like the spread of a species or the fluctuations of the stock market.

Consider a species invading a new habitat. Individuals disperse from their birthplace, move some distance, and reproduce. The distribution of dispersal distances is called a ​​dispersal kernel​​. If this kernel is thin-tailed (e.g., Gaussian), individuals tend to stay close to home. The population spreads like a wave with a constant speed. The front advances steadily, "pulled" by the individuals at the leading edge.

But what if the dispersal kernel is heavy-tailed? This means there's a non-trivial chance that an individual makes a huge leap, far ahead of the established front. This single long-distance founder can establish a new, distant colony. This colony grows and, in turn, sends out its own long-distance dispersers. The result is that the overall rate of spread is no longer constant; it is driven by these rare, extreme leaps. The invasion front ​​accelerates​​ over time. This principle is not only key to understanding biological invasions but also to the persistence of metapopulations, where long-distance colonization can rescue distant patches from extinction.

This focus on the "largest leap" brings us to the ultimate question of extremes: What is the biggest event we can expect? Again, our intuition, trained by the Central Limit Theorem for sums, fails us. For maxima, we need a different law: the ​​Fisher-Tippett-Gnedenko theorem​​, a cornerstone of ​​Extreme Value Theory​​. It states that if you take the maximum of a large number of random variables, its distribution (after suitable normalization) will converge to one of only three possible types.

  • ​​Type I (Gumbel):​​ For parent distributions with thin, exponential-like tails (like the Normal distribution).
  • ​​Type III (Weibull):​​ For parent distributions with a strict upper limit (e.g., human height).
  • ​​Type II (Fréchet):​​ For heavy-tailed parent distributions with power-law decay.

This is profound. If you are modeling daily returns on a speculative cryptocurrency whose distribution exhibits a power-law tail, the largest single-day crash (or rally) over a period of years will not be described by a Gaussian. It will follow a Fréchet distribution. This is the mathematically correct framework for risk management, because it was built specifically to describe the behavior of the "black swan" events that dominate the heavy-tailed world.

The Convergence Trap: On Being Almost Surely Right, but Wrong on Average

Perhaps the most subtle and dangerous aspect of heavy tails is how they can trick us into thinking our methods are working when they are not. This is the convergence trap.

In many simulations, we rely on the Law of Large Numbers, which states that the average of our samples will converge to the true expected value. We might check this by seeing if our simulation result gets closer and closer to the right answer. But there are different kinds of convergence.

Imagine a bizarre numerical scheme designed to approximate a value that is known to be zero. Let's say the scheme's output after nnn steps, X(n)X^{(n)}X(n), is given by the expression n×1{Y>n}n \times \mathbf{1}_{\{Y>n\}}n×1{Y>n}​, where YYY is a random variable drawn from a particularly heavy-tailed Pareto distribution. For any single run of the simulation, the value of YYY is some fixed, finite number, say yruny_{run}yrun​. As we increase nnn, eventually nnn will become larger than yruny_{run}yrun​. For all subsequent steps, the condition {Y>n}\{Y>n\}{Y>n} will be false, the indicator function 1{Y>n}\mathbf{1}_{\{Y>n\}}1{Y>n}​ will be zero, and our approximation X(n)X^{(n)}X(n) will be exactly zero. So, for any given run, our approximation eventually converges perfectly to the right answer. This is called ​​almost sure convergence​​. It seems like our method is a success.

Now, let's look at the average error. The error is just X(n)X^{(n)}X(n) itself. Its expectation is E[X(n)]=E[n×1{Y>n}]=n×P(Y>n)\mathbb{E}[X^{(n)}] = \mathbb{E}[n \times \mathbf{1}_{\{Y>n\}}] = n \times \mathbb{P}(Y>n)E[X(n)]=E[n×1{Y>n}​]=n×P(Y>n). For the specific Pareto distribution used in this thought experiment, it turns out that P(Y>n)=1/n\mathbb{P}(Y>n) = 1/nP(Y>n)=1/n. So the expected error is n×(1/n)=1n \times (1/n) = 1n×(1/n)=1. The average error never goes to zero. It stays stubbornly at 1, no matter how large nnn gets.

What is happening? As nnn increases, the probability of the error being non-zero, P(Y>n)\mathbb{P}(Y>n)P(Y>n), shrinks. But on those increasingly rare occasions when the error is non-zero, its magnitude, nnn, grows just as fast. The shrinking probability is perfectly cancelled by the growing magnitude of the error. The average remains constant. This is a failure to converge "in mean" (L1L^1L1). It's a situation where the interchange of limits and expectations, a step we often take for granted, is forbidden.

This trap appears in many real-world simulations. When we use Monte Carlo methods to calculate free energies in chemistry, the "importance weights" we compute can have a heavy-tailed, or even infinite-variance, distribution. Even though our estimate might seem to be settling down, its variance can be so huge that the number of samples required for a reliable answer is astronomically large. Similarly, when we try to generate random numbers from a heavy-tailed distribution using methods like inverse transform sampling, the problem becomes numerically ​​ill-conditioned​​. A tiny floating-point error in our input uniform random number uuu near 1 can be amplified into a colossal error in the output, requiring careful, specialized algorithms to manage. Standard statistical tools, like confidence intervals, can also fail, providing a false sense of security by under-reporting the true uncertainty when applied to heavy-tailed data.

The world of heavy tails is a fascinating and counter-intuitive place. It's a reminder that the "average" is not always the most important story, and that the single, dramatic outlier can sometimes have more to say than all the well-behaved data points combined. It forces us to be more humble in our statistics, more robust in our methods, and more alive to the possibility of the extreme. From the journey of a single enzyme to the invasion of a continent, from the glow of an atom to the crash of a market, the mathematics of heavy tails provides a unified language to describe the magnificent and sometimes terrifying power of the outlier.

Applications and Interdisciplinary Connections

Now that we have grappled with the principles of heavy-tailed distributions—their power-law falloffs, their often-infinite moments, and the strange dominance of the single largest event—we can embark on a journey. We are going to see that this is not some esoteric corner of statistics. It is a unifying concept that appears with startling regularity across the vast landscape of science and engineering. From the lottery of life in a single bacterium to the architecture of the human brain, from the performance of our global computer networks to the quest for limitless energy, the signature of the heavy tail is everywhere. It is a clue, left by nature, that reveals a deep and often counter-intuitive structure in the world.

The Imprint of Life and Nature

Our story begins with one of the most fundamental questions in biology: how does evolution work? In the 1940s, Luria and Delbrück devised a brilliant experiment to determine if bacterial resistance to a virus was a mutation acquired in response to the threat, or if it arose spontaneously and randomly before the encounter. If resistance is acquired on the spot, every bacterium has a small, equal chance of surviving, and the number of survivors across different cultures should follow a well-behaved Poisson distribution, where the variance is equal to the mean. But if mutations arise spontaneously during population growth, an early mutation creates a huge "jackpot"—a large clone of resistant descendants. A late mutation creates only a tiny one. This "rich-get-richer" dynamic of clonal expansion results in a distribution of survivors with a variance far, far larger than its mean. The data showed exactly this wild dispersion, a Fano factor F=Var(M)/E[M]≫1F = \mathrm{Var}(M)/\mathbb{E}[M] \gg 1F=Var(M)/E[M]≫1, providing definitive evidence for spontaneous mutation. This was one of the first times a heavy-tailed statistical signature was used to uncover a fundamental mechanism of life.

This same "jackpot" principle scales from the microscopic to the macroscopic. Consider the patterns of forest fires. Many fires are small and fizzle out quickly. But a rare few, under just the right conditions of wind, fuel, and dryness, explode into catastrophic conflagrations that account for the vast majority of total area burned. The distribution of fire sizes, like the distribution of bacterial mutants, is profoundly heavy-tailed. Understanding this is not merely academic; it is critical for managing risk. Modern ecological models use frameworks like Bayesian statistics to analyze historical fire data, fitting them to distributions like the Pareto. By doing so, they can move beyond simple averages and begin to quantify the probability of the next catastrophic "jackpot" event, providing a principled way to prepare for the extremes that inevitably shape our ecosystems.

Perhaps most astonishingly, this pattern may be etched into the very structure of our minds. Neuroscientists mapping the intricate wiring diagrams of the brain—the connectome—often analyze them as networks. A key question is about the distribution of connections per neuron, or the "degree distribution." In many systems, from the simple nematode C. elegans to the vastly more complex mouse brain, this distribution appears to be heavy-tailed. There are many neurons with few connections, and a few "hub" neurons with an enormous number of connections. This hints at a "scale-free" or similar architecture, which is theorized to be robust and efficient at processing information. Of course, biology is messy. Unlike clean mathematical models, real brain data often shows that while the degree distribution is heavy-tailed, it may not be a perfect power-law. It could be a truncated power-law or a log-normal distribution. The scientific debate is active, but the central point remains: the concept of heavy-tailed distributions provides the essential language and tools for investigating the fundamental design principles of the brain.

The Ghost in the Machine

It is a curious fact that the statistical patterns we discover in nature are often unintentionally recreated in the technological world we build. Our digital universe is haunted by the same heavy-tailed ghosts.

Consider the simple act of reading a file from an old spinning hard disk. If the file's data blocks are scattered across the disk in a linked chain, the disk head must perform a "seek"—a physical movement—to find each subsequent block. Most seeks are quick, but occasionally, the head must travel all the way across the disk platter, resulting in a seek time that is orders of magnitude longer than the average. This distribution of seek times is heavy-tailed. The total time to read the file is the sum of all these individual seek times. Here, the "single large jump" principle becomes vividly apparent: the total latency is utterly dominated by the single slowest seek. Even if a file has thousands of blocks, one unlucky, long-distance journey of the disk head can make the entire read operation feel grindingly slow. The tail of the total latency distribution inherits its character directly from the tail of the single-seek distribution, providing a stark lesson in system bottlenecks.

This principle scales up from a single component to massive, distributed systems. Imagine a data center with many servers trying to balance an incoming stream of computing jobs. Some jobs are small and finish quickly (light-tailed), while others are enormous and can run for hours (heavy-tailed). A common-sense approach to load balancing might be to distribute all jobs randomly across all servers, to "average out" the load. Heavy-tail theory reveals this intuition to be catastrophically wrong. When you spread the "poison" of heavy-tailed jobs everywhere, every server queue is at risk of being blocked by a monster job. A tiny, urgent query can get stuck behind a massive data-crunching task, leading to terrible response times for everyone. A far better, if counter-intuitive, strategy is isolation: create a dedicated pool of servers just for the heavy jobs, and let the vast majority of light jobs run freely on their own protected servers. By containing the rare, extreme events, you improve the performance for the common case, dramatically reducing the high-percentile waiting times that users actually experience.

The influence of heavy tails extends even to the content that flows through our machines. The very reason we can compress an image of a natural scene into a JPEG file is a consequence of sparsity, which is the twin sister of heavy-tailedness. When an image is passed through a mathematical prism like a wavelet transform, it is broken down into coefficients that represent features at different scales and orientations. It turns out that for natural images, the histogram of these coefficients is classicly heavy-tailed: a huge number of coefficients are nearly zero, but a tiny fraction are very large. These few large coefficients capture almost all the important visual information—the edges, the textures, the contours. Compression algorithms work by mercilessly throwing away the sea of near-zero coefficients and carefully preserving the few large ones. The high kurtosis (κ≫3\kappa \gg 3κ≫3) of this distribution is the statistical signature of sparsity, and it is this structure in the world around us that makes digital representation and communication possible.

Taming the Beast in Data Science

If our data is populated by these wild, heavy-tailed distributions, how do we possibly make sense of it? Many of the workhorse algorithms of statistics and machine learning were built on the assumption of well-behaved, Gaussian-like data, and they can fail spectacularly in the face of heavy tails.

Take the k-means clustering algorithm, a standard tool for finding groups in data. The algorithm works by calculating the "center" (mean) of each cluster and minimizing the sum of squared distances to that center. Now, imagine your data is drawn from a distribution with infinite variance, as is the case for a power-law with exponent α∈(1,2)\alpha \in (1,2)α∈(1,2). The sample mean is no longer a stable estimator of the center; it's violently pulled around by the most extreme points in your sample. And because the algorithm squares distances, these outliers have an absurdly disproportionate influence on the outcome. The result is chaos. The algorithm ends up creating meaningless clusters just to isolate the few extreme points, and the final result is unstable and completely dependent on the initial random starting conditions. It's like trying to find the average position of a group of people when one of them is on the moon.

So, what can be done? We have two general strategies for taming this beast. The first is to design robust procedures that are less sensitive to extremes. When we evaluate a machine learning model using k-fold cross-validation, our estimate of the model's error is essentially an average of the errors on different data subsets. If the error distribution is heavy-tailed, a few data points on which the model fails spectacularly can completely dominate this average, giving a highly volatile and unreliable estimate of its true performance. A robust solution is to use a trimmed mean or a winsorized mean of the errors. By either discarding or capping the most extreme error values, we can obtain a much more stable and reliable estimate of the model's typical performance, without being misled by rare catastrophes.

The second, and often more powerful, strategy is to transform the data itself. If the magnitudes of our features are causing problems, we can simply throw them away while preserving what truly matters: the rank ordering. This is the idea behind quantile normalization. For each feature, we rank the data points from smallest to largest. Then, we replace each data point's original value with a new value drawn from a standard normal distribution, based on its rank. The lowest-ranked point gets a value from the low tail of the Gaussian, the median point gets a value of zero, and the highest-ranked point gets a value from the high tail. This transformation forces every feature into a well-behaved Gaussian shape, effectively neutralizing any outliers. This allows classical methods like Pearson correlation, which are highly sensitive to outliers, to function correctly again, leading to much more stable and reliable feature selection.

Frontiers of Discovery

The search for heavy-tailed phenomena continues at the very frontiers of science. Inside a tokamak fusion reactor, where we attempt to bottle a star to harness its energy, the flow of heat from the scorching hot core to the cooler edge is not a simple, steady stream. It is governed by violent turbulence, which often manifests as intermittent "avalanches" of heat that burst outwards. Simple models of transport, which assume heat diffuses locally like in a metal bar, fail to capture this behavior.

How do we detect such complex, nonlocal events in the roiling chaos of a plasma? Once again, we turn to statistics. By measuring the heat flux qqq and the temperature gradient ∇T\nabla T∇T over time, physicists can analyze their relationship. If the transport were simple diffusion, the fluctuations of the flux around its mean value should be Gaussian. A key diagnostic is the conditional kurtosis—the "tailedness" of the flux fluctuations for a given temperature gradient. An observation that the kurtosis is systematically much greater than 3 (K≫3K \gg 3K≫3) is a smoking gun. It provides quantitative evidence that the transport is not simple diffusion, but is instead dominated by intermittent, bursty events, pointing towards a deeper, nonlocal physics of avalanching that must be understood to control the plasma.

From biology to physics, from nature to technology, heavy-tailed distributions emerge as a surprisingly universal theme. They are the statistical fingerprint of a wide class of generative processes—systems with feedback, "rich-get-richer" dynamics, multiplicative growth, or systems poised at a state of self-organized criticality. Seeing this single mathematical idea illuminate so many disparate corners of the world is a testament to the profound unity of scientific principles. It reminds us that by listening carefully to the patterns in our data, we can learn the fundamental rules of the game.