try ai
Popular Science
Edit
Share
Feedback
  • Likelihood-Free Inference

Likelihood-Free Inference

SciencePediaSciencePedia
Key Takeaways
  • Likelihood-Free Inference (LFI) enables Bayesian analysis for complex simulation-based models where the likelihood function cannot be calculated.
  • Approximate Bayesian Computation (ABC) is an intuitive LFI method that approximates the posterior by accepting parameters from simulations that produce data similar to observed data.
  • Modern neural methods, such as Neural Posterior Estimation (NPE), learn the entire posterior distribution, offering vast improvements in computational efficiency.
  • Validation is critical; Simulation-Based Calibration (SBC) checks the inference algorithm, while Posterior Predictive Checks (PPC) assess the model's fit to reality.
  • LFI acts as a universal bridge connecting complex models to data, with transformative applications in fields ranging from population genetics to cosmology.

Introduction

At the heart of scientific discovery is our ability to update our beliefs in light of new evidence, a process formally described by Bayesian inference. This framework hinges on a crucial component: the likelihood function, which quantifies the probability of observing our data given a specific hypothesis. But what happens when our models of the world—simulations of galaxy formation, viral outbreaks, or particle collisions—become so complex that this likelihood is impossible to write down? This is the central challenge in much of modern science, where we can simulate a world but cannot calculate the probability of its existence.

This article addresses this critical gap by exploring the powerful world of Likelihood-Free Inference (LFI), a suite of methods designed to perform principled statistical inference using only a simulator. We will unpack how scientists can reason about model parameters even when the mathematical bedrock of traditional Bayesian analysis is missing. You will learn how these techniques build a robust bridge between our most ambitious theories and the real-world data we collect.

First, in "Principles and Mechanisms," we will dissect the core concepts of LFI, starting with the intuitive idea of Approximate Bayesian Computation (ABC) and the art of choosing informative summary statistics. We will then journey to the cutting edge, exploring how the machine learning revolution has given rise to highly efficient neural network-based approaches. Following this, the "Applications and Interdisciplinary Connections" section will showcase LFI in action, revealing how this single framework unlocks insights across a stunning diversity of fields, from decoding the blueprints of life in our DNA to measuring the fundamental parameters of our universe.

Principles and Mechanisms

The Heart of the Problem: When the Likelihood is Lost

At the very core of modern scientific inference lies a beautifully simple relationship first articulated by the Reverend Thomas Bayes more than two centuries ago. It tells us how to update our beliefs in the face of new evidence. In its modern form, we write it like this:

p(θ∣x)∝p(x∣θ) p(θ)p(\boldsymbol{\theta} | x) \propto p(x | \boldsymbol{\theta}) \, p(\boldsymbol{\theta})p(θ∣x)∝p(x∣θ)p(θ)

Let's take a moment to appreciate what this says. On the right, we have p(θ)p(\boldsymbol{\theta})p(θ), our ​​prior​​ belief about the parameters θ\boldsymbol{\theta}θ that govern our world. These parameters could be anything from the mass of a fundamental particle to the transmission rate of a virus. Next to it is the term p(x∣θ)p(x | \boldsymbol{\theta})p(x∣θ), the ​​likelihood​​. This is the star of the show. It answers the question: "If the true parameters of the universe were θ\boldsymbol{\theta}θ, what would be the probability of observing the specific data xxx that we just collected?" By multiplying our prior beliefs by the likelihood of the evidence, we arrive at the left side, p(θ∣x)p(\boldsymbol{\theta} | x)p(θ∣x), the ​​posterior​​ distribution—our updated, refined belief about the parameters after seeing the data.

For centuries, this formula has been the bedrock of statistics. But what happens when our model of the world becomes so complex, so intricate, that we can no longer write down the likelihood function p(x∣θ)p(x | \boldsymbol{\theta})p(x∣θ)?

This isn't a rare or academic problem. It is arguably the central challenge in much of 21st-century science. Consider the Large Hadron Collider. A physicist might have a theory with certain parameters θ\boldsymbol{\theta}θ, like the strength of a new force. To connect this theory to data, they must simulate what happens when two protons collide: a cascade of quarks and gluons, which form into a shower of particles, which then interact with a detector of baffling complexity. The end result is a data pattern xxx. The process involves so many layers of randomness and intractable physics that writing an explicit formula for p(x∣θ)p(x | \boldsymbol{\theta})p(x∣θ) is simply impossible. All we have is a simulator—a computer program that, given θ\boldsymbol{\theta}θ, can generate a synthetic data pattern xxx. We can create worlds, but we cannot calculate the probability of any single world.

The situation can be even more profound. Sometimes, the very concept of a likelihood density breaks down. Imagine a cosmological simulator that, for a given set of parameters θ\boldsymbol{\theta}θ, generates points xxx in a three-dimensional space. But suppose the physics of the model are such that the outputs must always lie on a thin, curved one-dimensional string within that space. The probability of the simulator producing an output that is exactly a particular point xxx on the string is zero, and the probability of it producing an output anywhere off the string is also zero. A smooth probability density function, the kind we can write down and evaluate, simply does not exist. We are left with a simulator that we can run forward, but whose likelihood we cannot grasp.

This is the frontier. We have powerful, mechanistic models of reality—agent-based models of infections, simulations of nuclear reactors, models of galaxy formation—but we cannot use Bayes' rule in the traditional way. We are adrift without a likelihood. So, what can we do?

A Child's Game of "Getting Warmer": Approximate Bayesian Computation

When a powerful formula breaks, the most brilliant solutions are often stunningly simple. The first and most intuitive approach to likelihood-free inference is a method called ​​Approximate Bayesian Computation​​, or ​​ABC​​. It's best understood as a game of "hot or cold."

Imagine you want to find the parameters θ\boldsymbol{\theta}θ that describe our universe. You have your simulator, which can create synthetic universes. The game is this:

  1. Pick a set of parameters θ∗\boldsymbol{\theta}^*θ∗ at random from your prior distribution of beliefs, p(θ)p(\boldsymbol{\theta})p(θ).
  2. Run your simulator with these parameters to generate a synthetic dataset, xsimx_{sim}xsim​.
  3. Compare your synthetic data xsimx_{sim}xsim​ to your real, observed data xobsx_{obs}xobs​.
  4. If they look "close enough," you declare you're "warm" and you keep the parameters θ∗\boldsymbol{\theta}^*θ∗. If they don't, you're "cold," and you throw those parameters away.

If you repeat this game millions of times, the collection of "warm" parameters you've kept will form an approximation to the true posterior distribution p(θ∣xobs)p(\boldsymbol{\theta} | x_{obs})p(θ∣xobs​). Why? Because you've systematically filtered for parameters that are capable of producing a world that looks like ours.

Of course, we have to be more precise about "close enough." We need to define two things: a ​​distance function​​, ρ(x1,x2)\rho(x_1, x_2)ρ(x1​,x2​), to measure how far apart two datasets are, and a ​​tolerance​​, ϵ\epsilonϵ, that defines our circle of acceptance. The formal rule becomes: accept θ∗\boldsymbol{\theta}^*θ∗ if ρ(xsim,xobs)≤ϵ\rho(x_{sim}, x_{obs}) \le \epsilonρ(xsim​,xobs​)≤ϵ. The set of accepted parameters is then a sample from an approximate posterior, pϵ(θ∣xobs)p_{\epsilon}(\boldsymbol{\theta} | x_{obs})pϵ​(θ∣xobs​).

This simple rejection algorithm is the historical heart of ABC. But it immediately runs into a new problem. For a high-dimensional dataset like the output of a particle detector or a snapshot of a galaxy, the probability of a simulated dataset being "close" to the observed one in every single dimension is vanishingly small. This is the infamous "curse of dimensionality." If we try to match the entire dataset, our acceptance rate will be so close to zero that we would need to simulate for longer than the age of the universe to get a decent sample.

The solution is another act of scientific simplification: we don't compare the entire datasets. We compare a handful of intelligently chosen ​​summary statistics​​.

The Art of Abstraction: Choosing What Matters

Instead of asking if the entire simulated universe xsimx_{sim}xsim​ looks like our observed universe xobsx_{obs}xobs​, we ask if a small vector of summary statistics s(xsim)s(x_{sim})s(xsim​) looks like the summaries of our observed universe s(xobs)s(x_{obs})s(xobs​). Our acceptance rule becomes ρ(s(xsim),s(xobs))≤ϵ\rho(s(x_{sim}), s(x_{obs})) \le \epsilonρ(s(xsim​),s(xobs​))≤ϵ. The whole game now hinges on our ability to choose good summaries. What makes a summary "good"?

In an ideal world, we would find a ​​sufficient statistic​​. This is a magical summary that captures all the information about the parameters θ\boldsymbol{\theta}θ that was contained in the original, high-dimensional data. If we use a sufficient statistic, we lose absolutely nothing in the compression. In the limit as our tolerance ϵ\epsilonϵ goes to zero, our ABC procedure will converge to the exact, true posterior distribution.

In the real world of complex models, however, sufficient statistics are almost never available. We have to use our scientific intuition to hand-craft summaries that are highly informative about the parameters we care about. This is where domain knowledge becomes indispensable. Consider an agent-based model of a bacterial infection, where we want to infer three parameters: the bacterial replication rate λ\lambdaλ, the neutrophil killing rate μ\muμ, and the neutrophils' chemotactic sensitivity χ\chiχ. A brilliant choice of summaries would be:

  • To learn about λ\lambdaλ, we can measure the slope of the logarithm of the bacterial population during the initial phase of infection, when growth is nearly exponential.
  • To learn about μ\muμ, we can focus on moments when neutrophils and bacteria are in contact and measure the effective per-contact killing hazard.
  • To learn about χ\chiχ, we can measure how well neutrophil velocity vectors align with chemical gradients (a "chemotactic index") or how tightly they cluster around bacterial colonies (a "pair-correlation function").

Each summary is mechanistically linked to a specific parameter. They are designed to "disentangle" the effects of the different parameters, preserving the identifiability of each one.

But even with clever summaries, there's another layer of subtlety: how do we define the distance ρ\rhoρ? If we have three summaries, and one of them is naturally much noisier (has a larger variance) than the others, a simple Euclidean distance will be dominated by this noisy component. We might end up accepting simulations that match the noise but fail to match the more informative signals.

The solution is to use a more intelligent metric. We can standardize the summaries by dividing by their standard deviations. Even better, we can use a ​​Mahalanobis distance​​. This distance accounts not only for the different scales (variances) of the summaries but also for the correlations between them. It's like finding the perfect coordinate system in which to measure distance, down-weighting noisy and redundant directions to focus on what truly matters for telling different parameter values apart.

Beyond Brute Force: The Neural Revolution

ABC is intuitive and powerful, but its brute-force rejection of simulations is wasteful. It's not uncommon to run a million simulations and accept only a few hundred. For the last decade, scientists have been asking: can we do better? Can we learn from all the simulations, not just the ones that happen to land inside our tiny acceptance region?

The answer, driven by the revolution in machine learning, is a resounding yes. This has led to a new generation of LFI methods that are orders of magnitude more efficient.

One of the most powerful is ​​Neural Posterior Estimation (NPE)​​. Instead of just trying to collect samples from the posterior, NPE uses a deep neural network to learn the entire posterior distribution p(θ∣s(x))p(\boldsymbol{\theta} | s(x))p(θ∣s(x)) as a function. The training process is simple in concept:

  1. Generate a large number of simulated data summaries sis_isi​ from parameters θi\boldsymbol{\theta}_iθi​ drawn from the prior. This gives you a training set of pairs (θi,si)(\boldsymbol{\theta}_i, s_i)(θi​,si​).
  2. Train a flexible conditional density estimator (like a normalizing flow) to learn the mapping. You are essentially teaching the network: "When you see an input summary that looks like sis_isi​, output a probability distribution that is sharply peaked around θi\boldsymbol{\theta}_iθi​."

Once trained, the neural network becomes a reusable inference machine. You feed it your single, real observed summary statistic s(xobs)s(x_{obs})s(xobs​), and it instantly returns a fully-fledged, analytical approximation of your posterior distribution. This property is called ​​amortization​​: the heavy computational cost of simulation is paid once, upfront, during training. Afterwards, inference for any new observation is incredibly fast. This is a monumental leap in efficiency compared to running a full ABC analysis from scratch for every new dataset.

Other neural methods learn different parts of the Bayesian equation. Neural Likelihood Estimation (NLE) learns the likelihood function p(s(x)∣θ)p(s(x) | \boldsymbol{\theta})p(s(x)∣θ), while Neural Ratio Estimation (NRE) learns the ratio of likelihoods for different parameters—a quantity that is often all that is needed for inference. Together, these methods represent a paradigm shift, turning the difficult problem of inference into a machine learning task.

Are We Even Right? Calibration and Model Checking

We have developed these wonderfully sophisticated tools. But with great power comes great responsibility. How do we know our fancy neural network, or even our simple ABC algorithm, is giving us a correct answer? And even if the algorithm is correct, how do we know our underlying simulator is a good model of reality? These are two distinct, and equally vital, questions.

To answer the first question—Is my inference machine working correctly?—we use a procedure called ​​Simulation-Based Calibration (SBC)​​. SBC is an internal consistency check. We play a game against ourselves. We generate a "ground truth" parameter θtrue\boldsymbol{\theta}_{true}θtrue​ from our prior, run our simulator to get a fake observation y~\tilde{y}y~​, and then run this fake observation through our entire inference pipeline to get a posterior. We then check where θtrue\boldsymbol{\theta}_{true}θtrue​ falls within our calculated posterior. If our inference machine is well-calibrated, then over many repetitions of this game, the "true" parameter should fall in the bottom 10% of our posterior 10% of the time, between the 10th and 20th percentile 10% of the time, and so on. The distribution of these "ranks" should be perfectly uniform. If it's not, our machine is biased. SBC is the gold standard for debugging our inference algorithm.

But SBC only tells us that our tool is working correctly under the assumption that our simulator is the true model of the world. It's an internal check. To answer the second question—Is my model a good description of reality?—we need an external check against the real data. This is the job of ​​Posterior Predictive Checks (PPC)​​. The process is:

  1. Run your inference on the real data xobsx_{obs}xobs​ to get your posterior distribution for the parameters.
  2. Draw many parameter sets from this posterior.
  3. For each parameter set, run your simulator to create an whole ensemble of replicated universes, xrepx_{rep}xrep​.
  4. Now, compare your single real universe xobsx_{obs}xobs​ to this ensemble. Does it look like a typical member? Or is it a bizarre outlier that your model, even with its best-fit parameters, can't seem to replicate?

If the real data looks strange compared to the posterior predictions, it's a red flag that your model is fundamentally wrong. It's missing some crucial physics, biology, or economics. Together, SBC and PPC allow us to be responsible scientists: SBC validates our tools, and PPC validates our theories.

The Beauty of Being Wrong: Inference Under Misspecification

So what happens when our PPCs fail, and we are forced to admit that our model is wrong? As the statistician George Box famously said, "All models are wrong, but some are useful." Is inference pointless if our simulator isn't a perfect replica of reality?

The beautiful answer is no. Likelihood-free inference is more robust and graceful than one might think. When we try to fit a misspecified model, the inference process doesn't simply break. Instead, it finds the parameter value within our assumed model family that is closest to the true data-generating process, in a precise information-theoretic sense (minimizing the Kullback-Leibler divergence).

Imagine the true process generating our data is log-normal, but our simulator is only capable of producing Gaussian distributions. If we run our LFI pipeline, we won't get nonsense. The posterior will converge on the parameters of a very specific Gaussian: the one that has the exact same mean and variance as the true log-normal distribution.

This is a profound result. Even with a "wrong" model, we can learn correct and useful things about the world—in this case, its true mean and variance. The process of inference doesn't demand perfection from our models. It simply finds the best possible approximation to reality within the language of the model we have provided. It is this robustness that allows science to progress, step by step, building and refining imperfect models that, nevertheless, bring us ever closer to the truth.

Applications and Interdisciplinary Connections

Now that we have acquainted ourselves with the principles of likelihood-free inference, let us embark on a journey across the vast landscape of modern science. We will see how this single, powerful idea acts as a master key, unlocking secrets in realms as disparate as the intricate dance of our own genes and the majestic evolution of the cosmos. The story of science in the 21st century is increasingly told not with simple, elegant equations, but with complex, sprawling computer simulations—digital worlds that are often as rich and messy as reality itself. The grand challenge has been to build a bridge between these intricate simulations and the noisy, incomplete data we collect from the real world. Likelihood-free inference is that bridge. It is a universal translator, allowing us to hold a meaningful dialogue with our most ambitious models of reality.

Decoding the Blueprints of Life

Our journey begins with the very essence of life: our DNA. For over a century, we have understood that evolution proceeds by natural selection, but only recently have we gained the tools to read the story of that selection in the genome. Imagine trying to reconstruct the history of a population—its triumphs and struggles—from the subtle patterns of variation in the DNA of its descendants. Population geneticists build sophisticated computer programs, such as coalescent simulators, that model how genes mutate, recombine, and are passed down through generations under the influence of selection.

Suppose we identify a region in the human genome that appears to have been recently shaped by strong positive selection—a "selective sweep." We can now ask our simulation a very specific question: "What strength of selection, sss, and what timing of the event, τ\tauτ, would be most likely to produce the patterns of genetic diversity and linkage disequilibrium we observe today?" The likelihood function for this process is hopelessly complex. But with likelihood-free inference, we can simply run the simulation many times with different values of sss and τ\tauτ, and find which simulations produce genetic patterns that "look like" our data. This allows us to not only estimate the strength of ancient selection but even to distinguish between different modes of evolution, such as a "hard sweep" from a single new mutation versus a "soft sweep" from pre-existing variation.

From reading the blueprints of life, we move to writing them. In synthetic biology, scientists engineer novel gene circuits inside living cells. A common goal is to create a bistable "toggle switch," where a cell can exist in either an 'ON' or 'OFF' state. An experiment might reveal a population of cells with a bimodal distribution of a fluorescent reporter protein—some are dim, some are bright. But this observation is ambiguous. Does it represent true bistability, with individual cells capable of switching between states? Or does it simply reflect a heterogeneous population where some cells are permanently programmed to be 'low' and others 'high' due to cell-to-cell differences?

A static snapshot cannot tell the difference. The crucial evidence lies in the dynamics. By taking time-lapse movies of individual cells, we can search for the smoking gun: a cell spontaneously switching from 'OFF' to 'ON'. Likelihood-free inference provides a formal framework for this detective work. We can construct two competing models—one for true multistability (M1M_1M1​) and one for extrinsic heterogeneity (M0M_0M0​)—and simulate both. By choosing summary statistics that capture the dynamics of the system, such as the fraction of trajectories that exhibit a state transition or the average time spent in each state, we can ask which model's simulations best match the behavior of the real cells. This allows us to perform a rigorous Bayesian model selection, calculating the very probability that one explanation is correct over the other.

Modeling the Fabric of Society and Nature

The power of simulation-based inference truly shines when we study complex adaptive systems, where the behavior of the whole emerges from the interactions of many individual agents. Consider the spread of a healthcare-associated infection within a hospital. We can build an Agent-Based Model (ABM) where digital "agents" representing patients and clinicians move through interconnected wards, come into contact, and transmit pathogens based on stochastic rules, including the probability of hand hygiene compliance.

The exact path of any single outbreak is unpredictable, a unique historical accident. Thus, the likelihood of observing the exact time series of new cases in each ward is effectively zero. Yet, hospital administrators must make decisions. What level of hand-washing compliance is needed to control outbreaks? LFI allows us to connect our model to the data that matters. We don't ask the model to reproduce the exact history of the outbreak. Instead, we ask it to reproduce the statistical character of the outbreak. We choose summary statistics that capture the emergent, macroscopic features: the overall distribution of daily cases, the temporal correlation in case counts, the average time between generations of infection. By finding the model parameters that best reproduce these patterns, we can infer the underlying properties of the system, such as the hand hygiene compliance rate, and thereby guide real-world policy.

This same logic applies to the grand, chaotic systems of the natural world. Imagine looking at a satellite image of the aftermath of a massive wildfire. We see a complex, fractal-like burn scar etched into the landscape. This final shape is the result of a dynamic process driven by wind, fuel heterogeneity, and the spotting of new fires by flying embers. The likelihood of this specific scar shape arising, given the physics, is intractable. But with a good simulator of fire spread and LFI, we can play the role of a forensic scientist. We can propose different values for the parameters governing wind influence, fuel-driven spread, and ember spotting, and for each proposal, we can simulate a fire. By comparing the geometry of the simulated burn scars to the real one—using summary statistics that capture the scar's roughness, its anisotropy (is it elongated in the direction of the wind?), and its overall size—we can infer the physical conditions that gave rise to the observed event.

The Grand Design: From Microstructures to the Cosmos

One of the most profound aspects of physics is the way the same mathematical ideas recur at vastly different scales. It turns out that the tools used to characterize the structure of the entire universe can be adapted to understand the microscopic structure of a piece of metal. In cosmology, scientists study the "cosmic web," the vast network of galactic filaments and voids, using the tools of topology. They compute how the number of connected components (β0\beta_0β0​) and tunnels (β1\beta_1β1​) change as they threshold a density field. This method, known as persistent homology, produces a topological summary of the universe's structure.

Now, imagine a materials scientist creating a new alloy. The material's properties depend on its microstructure, formed by a process of nucleation and growth of crystal grains. By taking a microscope image, the scientist can see this structure. Using the exact same topological summary statistics—the Betti curves—they can characterize the geometry of the grains. By coupling a simulator of this nucleation-and-growth process with likelihood-free inference, they can infer the underlying physical parameter, such as the nucleation rate λ\lambdaλ, directly from the image. In a beautiful example of interdisciplinary exchange, a method forged for studying galaxies finds a home in a microscope.

Scaling up to cosmology proper, we find ourselves at the absolute cutting edge of LFI. Our standard model of the universe, Λ\LambdaΛCDM, is described by just a handful of parameters, such as the density of matter, Ωm\Omega_mΩm​, and the amplitude of fluctuations, σ8\sigma_8σ8​. We measure these by observing their effects on cosmic structures, for example, through the subtle distortion of distant galaxy images by weak gravitational lensing. The forward model involves simulating the evolution of the universe and the passage of light through it. The resulting likelihood function is notoriously complex due to non-linear physics, non-Gaussian fields, and the partial, masked nature of our sky surveys.

Here, even traditional ABC can be too slow. This has spurred the development of neural likelihood-free inference. We can train a powerful deep neural network to learn the mapping from an observed data summary (like the weak lensing power spectrum) directly to the posterior distribution of the cosmological parameters. In essence, the neural network becomes a highly efficient "inference engine," capable of analyzing new data in a fraction of a second, a task that would have taken millions of simulation runs with older methods. This remarkable synthesis of physics, statistics, and artificial intelligence is what allows us to wring precise knowledge from our most complex cosmological data.

Inner Space, Outer Space, and the Art of Asking Questions

The reach of LFI extends from the largest scales imaginable to the inner space of our own minds. Neuroscientists build neural mass models to understand brain rhythms, such as the pathological beta-band oscillations associated with Parkinson's disease. For simplified versions of these models—those that are linear and driven by Gaussian noise—an approximate likelihood can sometimes be written down. But the brain is not so simple. LFI liberates researchers to build models that are more faithful to the underlying biology, incorporating the strong nonlinearities and complex noise sources that are inherent to neural systems. It allows them to connect these more realistic simulations directly to electrophysiological data, like local field potentials recorded from deep within the brain. In high-energy physics, researchers use gargantuan simulators to model the intricate interactions of particles in detectors like the Large Hadron Collider. LFI provides a principled way to "unfold" the observed data, separating a faint physics signal from a large, time-varying background and thereby enabling joint inference on both the parameters of interest and the latent properties of the background itself.

Throughout this journey, a common theme has emerged: the critical importance of summary statistics. A likelihood-free inference is only as good as the question you ask of it, and the summary statistic is the question. If you want to infer the parameters of a flow through a porous medium, you cannot simply compare the entire, high-dimensional pressure field from your simulation to your sparse sensor measurements. The probability of an exact match is zero. This is the "curse of dimensionality." You must distill the essence of the observation into a low-dimensional summary—perhaps the effective conductivity or a few key moments of the pressure field. There is a fundamental trade-off: a summary that is too simple loses information and leads to a biased answer; a summary that is too complex makes the comparison computationally impossible. The design of summary statistics is where scientific intuition and statistical rigor meet, and it remains the central art of likelihood-free science.

Finally, what do we do when we have two competing conceptual models, and the likelihood is intractable for both? A hydrologist might have two different ideas about how a watershed responds to rainfall, both of which can be turned into simulators. Without a likelihood, traditional methods for model selection like the Bayes factor are out of reach. Here, we can appeal to a more fundamental principle of science: a good model should make good predictions. We can train each model on one part of our data and then see how well it predicts the data we held out. Using tools from decision theory, like strictly proper scoring rules, we can rigorously quantify which model provides more skillful forecasts. This pragmatic approach, focused on predictive performance, is a powerful and honest way to do science in the likelihood-free era.

From genes to galaxies, from epidemics to economies, simulation-based modeling is the common language of modern science. Likelihood-free inference is the grammar that allows us to speak it fluently, to ask meaningful questions of our most complex theories, and to learn from a world that is, and will always be, more wonderful and surprising than our equations alone can capture.