
In our quest to understand the cosmos, we face a fundamental limitation: we cannot run experiments on the universe. We have but one sky to observe, a single history to decipher. To bridge the gap between our theoretical models and the dazzling complexity of astronomical data, scientists create their own universes inside supercomputers. These digital cosmoses, known as mock catalogs, are indispensable tools that function as our cosmic laboratories. They allow us to test theories of galaxy formation, calibrate our observational techniques, and understand the inherent uncertainties in our measurements. Creating a universe that not only looks like our own but behaves like it is a profound challenge, blending physics, statistics, and computational science.
This article delves into the science and art of building and using these simulated universes. It addresses the critical question of how we translate the abstract laws of gravity and dark matter into a galaxy catalog that is statistically indistinguishable from what our telescopes see. Across the following chapters, you will gain a comprehensive understanding of this essential technique. The "Principles and Mechanisms" section will guide you through the construction process, from building a geometrically correct past lightcone and populating it with galaxies to meticulously mimicking the imperfections of a real survey. Following that, the "Applications and Interdisciplinary Connections" chapter will explore the vast scientific impact of mock catalogs, demonstrating how they are used to refine our measurements, test our theories of galaxy formation, and even probe the deepest mysteries of dark matter, gravity, and the fabric of spacetime itself.
To truly grasp our place in the cosmos, we need more than just a single snapshot of the universe. We need to understand its history, its dynamics, and the intricate web of physical laws that govern its evolution. Since we cannot rewind the cosmic clock or run experiments on real galaxies, we do the next best thing: we build our own universes inside supercomputers. These digital cosmoses, known as mock catalogs, are our laboratories for testing theories and understanding the subtleties of our own observations. But creating a universe that not only looks like our own but behaves like it is a profound scientific and artistic endeavor. It's a journey from the abstract laws of gravity to the dazzling tapestry of galaxies we see through our telescopes.
Imagine we have successfully run a massive cosmological simulation. Our computer has tracked the gravitational dance of billions of dark matter particles, forming a vast, filamentary network known as the cosmic web. The result is a cube of the universe, a snapshot in time showing the location of all matter at a single cosmic moment. But is this what a telescope sees?
Not at all. When we look out into the night sky, we are looking back in time. The light from the Andromeda Galaxy has traveled for 2.5 million years to reach us; we see it as it was 2.5 million years ago. The light from a distant quasar may have journeyed for 10 billion years. An astronomical survey, therefore, does not capture a 3D snapshot of the universe at a single time. Instead, it sees a collection of objects on our past lightcone—a 4D surface in spacetime where each object's distance is intrinsically linked to its lookback time.
To build a realistic mock catalog, we must abandon the simple 3D snapshot and construct this lightcone. We place a virtual observer at the center of our simulation box and ask: what would this observer see? As we look outwards in a certain direction, the farther we look, the earlier in the simulation's history we must draw our information.
This requires a precise mathematical recipe to connect an observable quantity, redshift (), to the comoving distance () in the simulation. Redshift, the stretching of light's wavelength due to cosmic expansion, is our primary indicator of distance. Starting from the fundamental principle that light travels along null geodesics in our expanding universe, one can derive a beautiful relationship:
Here, is the Hubble expansion rate at redshift , and is the speed of light. This integral tells us the total distance a photon has traveled to reach us from an object at redshift , carefully accounting for the fact that space itself was stretching during its entire journey. This equation is the geometric blueprint for our mock universe, allowing us to stack snapshots from different cosmic epochs to build a single, continuous lightcone that mirrors what a real telescope would see. This construction is not just a technical detail; it is the first principle of creating an honest representation of our observations, capturing the fact that the universe is not a static museum but an evolving entity. A lightcone mock naturally includes the evolution of galaxy populations, a feature entirely absent in a single snapshot.
Our simulation box is now geometrically correct, but it's still filled only with invisible dark matter. Where are the galaxies? Directly simulating the formation of individual stars and galaxies across cosmic volumes is computationally impossible. We need a cleverer, more statistical approach. This is where the Halo Model comes in.
The central idea is that all galaxies are born and live inside vast, invisible cocoons of dark matter called halos. The properties of a halo, primarily its mass, dictate the kinds of galaxies it can host. So, instead of building galaxies from scratch, we can write a recipe—a statistical prescription called the Halo Occupation Distribution (HOD)—that tells us how to populate the dark matter halos from our simulation with galaxies.
A standard HOD recipe distinguishes between two types of galaxies:
The full recipe for the mean number of galaxies in a halo of mass looks something like this:
Once we know how many satellites to place in a halo, we need to know where to place them. Here again, we turn to physics. Satellites should trace the gravitational potential of their host halo. The dark matter distribution in simulated halos is remarkably well-described by a universal formula, the Navarro-Frenk-White (NFW) profile. By drawing random positions for our satellite galaxies from this profile, we ensure they are distributed realistically within their host halos. The HOD is thus a powerful bridge, a simple yet physically motivated recipe that connects the invisible skeleton of the cosmos to the visible tracers we call galaxies.
Our synthetic universe is now filled with galaxies, correctly arranged on a lightcone. But we are not done yet. A real astronomical survey is an imperfect observer. It has blind spots, it can't see infinitely faint objects, and its measurements are noisy. For a mock catalog to be useful for science—for instance, to estimate the uncertainties in our measurements—it must be subjected to the same gauntlet of observational limitations as the real data.
First, a telescope can't look at the whole sky. Its view is limited by the Milky Way, bright stars, and the survey's chosen strategy. We model this by creating an angular survey mask, , a map on the sky that is 1 where the telescope observed and 0 where it didn't. Furthermore, even within the observed footprint, conditions aren't uniform. Atmospheric transparency, seeing, and instrumental noise can vary, affecting our ability to detect faint objects. This is captured by an angular completeness map, , which gives the probability of detection at each position.
Second, surveys are fundamentally flux-limited; we can only see objects brighter than some threshold. Since galaxies appear fainter with distance, this means our sample becomes progressively more biased towards intrinsically luminous galaxies at higher redshifts. This effect is described by a radial selection function, , which is the probability that a galaxy at redshift is included in our catalog. The number of galaxies we actually expect to see in a survey, , is a product of three factors: the intrinsic abundance of galaxies , the selection probability , and the volume of space in that redshift shell, :
This formula explains the characteristic shape of galaxy redshift distributions from surveys, which rise from zero, peak at some intermediate redshift, and fall off again at large distances where only the brightest objects are visible.
Finally, our measurements themselves are noisy. For many large surveys, measuring a galaxy's redshift precisely via spectroscopy is too time-consuming. Instead, astronomers estimate it from the galaxy's colors, a technique that yields a photometric redshift, or photo-. This is an inference problem: given a set of observed colors, what is the probability distribution of the true redshift? This process is not perfect. The resulting estimates have a characteristic scatter around the true value, a potential systematic bias, and, most troublingly, a fraction of catastrophic outliers where the estimate is wildly incorrect. A high-fidelity mock must model these photo- errors accurately.
The process of applying all these imperfections to our pristine simulated catalog is often done via a method called thinning. For each simulated galaxy, we calculate the total probability of it being observed, and then, with a roll of the dice, we decide whether to keep it or discard it. It's a process of systematic degradation, but it's essential for creating a mock that is statistically indistinguishable from the real data.
We have followed the recipe and cooked up a mock universe. How do we know it's any good? We must test it. This process, called validation, is not a single check but a hierarchical series of increasingly stringent tests, designed to ensure the mock behaves like the real universe on multiple levels.
The Basics: First, we check the most fundamental properties. Does the mock have the right number of galaxies as a function of redshift ()? Does it correctly reproduce the one-point distribution of galaxies, which tells us about the overall variance and non-Gaussianity of the cosmic field? These first steps ensure our modeling of the survey selection and basic galaxy abundance is correct.
Two-Point Clustering: Next, we ask if the galaxies are in the right places. We measure the two-point correlation function , which quantifies the excess probability of finding two galaxies separated by a distance . This is the fundamental measure of cosmic structure. We must check this both in 3D real space and in its 2D projection on the sky (), to validate that our line-of-sight projection and mask effects are correctly modeled.
Dynamics and Gravity: A static picture is not enough. We must check if our galaxies are moving correctly. Galaxies are constantly falling towards overdense regions due to gravity. This peculiar velocity adds to their cosmological redshift, creating a characteristic anisotropy in the observed clustering pattern known as Redshift-Space Distortions (RSD). By measuring the multipoles of the correlation function (the monopole, quadrupole, etc.), we can isolate this dynamical signature. A mock that fails to reproduce the observed RSD has an incorrect model of gravity or how galaxies trace the velocity field.
The Galaxy-Mass Connection: The ultimate test is to check if the mock correctly links visible galaxies to the underlying invisible matter. This is done using weak gravitational lensing. The gravity of foreground galaxies and their dark matter halos distorts the images of background galaxies. By measuring this coherent distortion, a signal called galaxy-galaxy lensing, we can directly weigh the matter around the foreground galaxies. A good mock must reproduce not only the clustering of galaxies (galaxy-galaxy correlation) and the lensing from all matter (shear-shear correlation), but also this crucial cross-correlation. It's the final verification that our HOD recipe has correctly captured the connection between light and mass.
Passing this entire hierarchy of tests gives us confidence in our mock. And we don't just need one. To understand the uncertainties in our cosmological measurements—how much our results might change if we could observe a different patch of the universe (an effect called cosmic variance)—we need an ensemble of hundreds or even thousands of mocks. A simple calculation shows that to estimate the variance of just a single measurement to 10% precision, we need about 200 independent mocks. This enormous computational demand is the driving force behind the development of ever-faster and more sophisticated mock generation techniques.
The story of mock catalogs is a story of ever-increasing fidelity, a constant push to incorporate more physics and confront new challenges. The "standard" mock described so far is already a magnificent achievement, but the frontiers of cosmology demand even more.
One major challenge is the effect of baryons (normal matter). Our dark-matter-only simulations ignore the complex physics of gas, star formation, and feedback from supernovae and active galactic nuclei (AGN). This feedback can expel gas from the centers of halos, making them less dense and suppressing the amount of structure on small scales. This is not a small effect; for weak lensing measurements, it can change the signal by 20% or more on the scales probed by current surveys. To account for this, modelers either modify the halo profiles in their HODs or apply a direct suppression factor to the matter power spectrum. Ensuring this is done in a way that is consistent for both galaxy clustering and weak lensing is a critical area of active research.
An even more fundamental frontier lies at the largest scales. Our simulations are based on Newtonian gravity, an excellent approximation within the cosmic web. But on scales approaching the size of the observable universe, the full machinery of General Relativity (GR) becomes important. Light's path and energy are affected not just by velocities, but by the gravitational potentials it traverses. Effects with exotic names like the Sachs-Wolfe, Integrated Sachs-Wolfe, and gravitational lensing magnification become significant, altering the observed number counts of galaxies in a way that Newtonian physics cannot capture. Modern mock catalogs are beginning to incorporate these effects, either by post-processing Newtonian simulations to calculate the relativistic potentials or by using new, cutting-edge relativistic N-body codes that solve Einstein's equations directly.
Finally, the sheer demand for vast numbers of mocks has sparked a revolution in methodology. Can we teach a machine to create a universe? Using techniques from artificial intelligence like Generative Adversarial Networks (GANs) and Variational Autoencoders (VAEs), researchers are training models to learn the mapping from cosmological parameters to galaxy catalogs directly from simulations. The challenge is to build these powerful models in a way that respects the fundamental laws of physics and produces catalogs that pass the rigorous validation hierarchy. This is the frontier where physics-based modeling meets data-driven inference, promising a future where we can generate entire virtual skies on demand, each one a faithful replica of our own magnificent, evolving cosmos.
Having journeyed through the principles of how we construct our cosmic simulations, we now arrive at the most exciting part of our exploration: what do we do with them? If a cosmological simulation is a grand stage, then a mock catalog is the dress rehearsal for the cosmic play. It is our chance to test our equipment, refine our lines, and anticipate the surprises the real Universe might have in store for us. The applications of these mock catalogs are as vast and varied as the cosmos itself, extending far beyond simply making pretty pictures. They are indispensable tools at the very heart of modern physics, allowing us to sharpen our understanding of the universe, from the behavior of a single galaxy to the fundamental laws of nature.
At its core, a mock catalog is an exercise in translation. Our most fundamental theories describe the behavior of dark matter, the invisible scaffolding upon which the luminous universe is built. Cosmological simulations, like the Particle-Mesh methods we've encountered, are incredibly good at predicting how gravity will sculpt this dark matter into a vast, interconnected network of filaments and dense knots called halos. But we cannot see dark matter directly. We see galaxies. So, how do we bridge this gap?
This is the first and most fundamental application of the mock catalog framework: to populate a simulated dark matter universe with galaxies. We develop a set of rules, a sort of cosmic recipe, to decide where galaxies should form and what properties they should have. A common approach, known as the Halo Occupation Distribution (HOD), uses a statistical model to answer questions like: for a dark matter halo of a certain mass, what is the probability that it hosts a central galaxy? And how many smaller "satellite" galaxies are likely to be orbiting within it? By applying these rules, we transform a map of dark matter into a vibrant catalog of galaxies, complete with positions and masses—a synthetic universe ready for study. This process is our first step in creating a meaningful link between theory and observation.
Once we have a universe in a box, what's its purpose? One of the most critical roles of a mock catalog is to act as a perfect "answer key" for testing the complex machinery of observational astronomy. When a real telescope, like the Vera C. Rubin Observatory, surveys the sky, it produces terabytes of raw data that must be processed by intricate software pipelines. These pipelines have to identify galaxies, measure their properties, and account for all sorts of observational artifacts: gaps in the data from bright stars, varying atmospheric conditions, and the subtle distortions of gravitational lensing. How do we know if this software is working correctly? How can we be sure it isn't introducing subtle biases that could lead us to the wrong cosmological conclusions?
We test it on a mock catalog. We can take our pristine, simulated galaxy catalog and "observe" it with a virtual telescope. This involves projecting the 3D catalog onto a 2D sphere, just as we see the sky, and applying all the known limitations of a real survey—the footprint on the sky, regions of poor data quality, and measurement noise. The result is a synthetic dataset that looks, for all intents and purposes, exactly like the real thing. But unlike the real Universe, we have the answer key. We know the true number of galaxies and their true properties. By comparing the output of our analysis pipeline with the known ground truth of the mock, we can validate our methods, hunt down bugs, and ensure our measurements of the real cosmos are as accurate as possible.
Furthermore, mock catalogs are essential for understanding the inherent uncertainties in our measurements. When we measure a property of the cosmos, like how galaxies cluster together, our measurement is subject to two primary sources of error. The first is "shot noise," which arises from the simple fact that galaxies are discrete points. It's like trying to gauge the average color of a sandy beach by looking at a handful of grains; the smaller your sample, the more your result will fluctuate. The second, more profound, source of error is "sample variance" or "cosmic variance." This is the uncertainty that comes from the fact that we only have one Universe to observe. Our particular patch of the cosmos might be slightly more or less dense than the cosmic average, simply by chance. How can we possibly estimate the uncertainty from this cosmic lottery?
We do it by simulating many mock universes. Each mock catalog is an independent realization of the cosmos, and by analyzing hundreds or thousands of them, we can measure the full range of possible outcomes. This allows us to build a precise statistical understanding of the uncertainties in our measurements. It's also why mock catalogs are indispensable for techniques like measuring the two-point correlation function, where a "random" catalog—a mock catalog with no clustering—is required to correct for the geometry of the survey and provide a baseline for what "no clustering" looks like.
The rules for painting galaxies onto dark matter halos are more than just a convenience; they are a direct probe of the physics of galaxy formation. A simple model might assume that the only thing that matters is the mass of the halo. But the Universe is more subtle. Imagine two dark matter halos of exactly the same mass. One might have formed peacefully and early in cosmic history, while the other grew violently through recent, major mergers. Is it not plausible that these different histories—this "assembly bias"—would affect the kinds of galaxies that form within them?
Mock catalogs are our laboratory for testing these ideas. We can build sophisticated models, such as Conditional Abundance Matching, where a galaxy's properties, like its color (a proxy for its age), are tied not just to the halo's mass but also to a secondary property that traces its formation history. We can then check if these more complex mocks produce galaxy populations that cluster in a way that better matches the real Universe.
This is not merely an academic detail. The stakes are immense. Our most precise cosmological measurements rely on using galaxies as tracers of the underlying cosmic structure. For instance, the Baryon Acoustic Oscillation (BAO) feature in the distribution of galaxies serves as a "standard ruler" to measure the expansion history of the universe. But what if our selection of galaxies for this measurement—say, by picking only old, red galaxies—is subtly biased by assembly effects? It could mean that our ruler is systematically warped, leading us to an incorrect measurement of the cosmic expansion rate [@problem_tca:3473083]. Mock catalogs that include these sophisticated physical effects are our only line of defense, allowing us to quantify and correct for these systematic errors, ensuring the integrity of our cosmological measurements.
The utility of mock catalogs extends far beyond galaxy surveys. They have become a universal tool for exploring a host of questions in fundamental physics, often in surprising and beautiful ways.
One of the deepest mysteries in physics is the identity of dark matter. Our standard model of Cold Dark Matter (CDM) predicts that large halos should be filled with a swarm of smaller subhalos, down to very small masses. Mock catalogs allow us to simulate what the signature of this substructure would be. In the phenomenon of strong gravitational lensing, a massive foreground galaxy can bend the light from a distant source into a spectacular ring or arc. If the lensing galaxy's halo is filled with invisible CDM subhalos, these subhalos will introduce tiny perturbations—wiggles and distortions—in the lensed image. By running mock lensing simulations, we can predict the exact form of these perturbations and design observational strategies to hunt for them. A detection—or a non-detection—of this predicted substructure would have profound implications for the nature of dark matter itself.
Is Einstein's theory of General Relativity the final word on gravity? Some theories attempting to explain the accelerated expansion of the cosmos propose modifications to GR. These theories must, however, look just like GR in dense environments like our solar system, or their effects would have already been detected. They achieve this through "screening mechanisms." One popular model, the "chameleon" mechanism, predicts that the modifications to gravity are suppressed in regions of deep gravitational potential. Mock catalogs provide a perfect tool to test this. We can take a mock catalog of galaxies, calculate the full gravitational potential at the location of each galaxy from all its neighbors, and identify which galaxies should be "screened" and which should be "unscreened". This allows us to make sharp predictions about where in the Universe to look for deviations from GR, turning the entire cosmos into a laboratory for fundamental physics.
Perhaps the most exciting new frontier is the application of mock catalog concepts to gravitational-wave astronomy. The ripples in spacetime from colliding black holes and neutron stars carry a wealth of information. By simulating catalogs of these merger events, we can develop and test the complex statistical tools needed to decipher their message. For example, the way a neutron star is tidally deformed just before being torn apart by a black hole depends sensitively on the equation of state (EOS) of matter at extreme nuclear densities. By creating mock catalogs of such events and running them through a Bayesian inference pipeline, we can perfect our ability to constrain the EOS and the population properties of black holes from real data, connecting cosmology with nuclear physics and extreme astrophysics.
We can even use these "standard sirens" to address the most foundational tenets of our cosmological model. The Cosmological Principle states that the Universe is homogeneous and isotropic on large scales—that it looks the same everywhere and in every direction. Is this true? With mock catalogs of standard sirens distributed across the sky, we can ask what kind of signal we would expect if the Hubble constant were not truly constant, but had a slight directional dependence. These simulations, which must carefully account for selection effects and observational uncertainties, allow us to design tests that could reveal a fundamental anisotropy in the cosmic expansion, a discovery that would reshape our entire understanding of the Universe.
From the humble task of placing galaxies in a simulation to testing the fabric of spacetime itself, mock catalogs have become the indispensable flight simulator for the modern cosmologist. They are the crucibles where our theories of galaxy formation are forged, the calibration tools that ensure the accuracy of our observations, and the testbeds for our explorations into the deepest questions of fundamental physics. They are, in essence, our way of asking the Universe, "What if?"