
In ecology, understanding where a species lives is a fundamental goal. Often, our primary source of information is presence-only data—a vast and growing collection of sightings from museum records, naturalists' logs, and modern citizen science apps. Each point confirms where a species has been found. The great challenge is to connect these dots to understand all the places a species could live. However, this data is not a pure reflection of biology; it is inherently skewed by where people happen to look, a problem known as sampling bias. This ghost in the machine can lead models to map observer behavior instead of species' habitats, creating a critical knowledge gap.
This article navigates the challenges and triumphs of using presence-only data. In the first chapter, "Principles and Mechanisms," we will dissect the problem of sampling bias and explore the elegant statistical strategies developed to account for it, clarifying what this data can and cannot tell us. Following that, in "Applications and Interdisciplinary Connections," we will witness how these corrected models are applied to paint maps of biodiversity, reconstruct the deep past, and forecast the future of life on a changing planet. We begin by examining the core mechanics of this data and the deceptive power of bias.
Imagine you are a detective trying to map the secret hideouts of a mysterious and elusive character. You don't have a complete list of their locations. Instead, you have a collection of scattered clues: a postcard sent from Paris, a receipt from a café in Cairo, a ticket stub from a theater in Tokyo. Each clue is a single, confirmed point of presence. This is the essence of presence-only data: a collection of dots on a map where a species has been seen. This data is abundant, flowing from centuries of museum collections, the logbooks of naturalists, and today, the smartphone apps of millions of citizen scientists. The grand ambition of ecologists is to connect these dots, to look at the environment at each location—the climate, the vegetation, the soil—and deduce a general rule, a "habitat profile," that describes all the places in the world the species could live. This is the work of Species Distribution Models (SDMs).
But right away, we hit a snag. The map of clues is not a true map of the character's preferences. It's a map of where they went and where a clue was left behind and found. Our collection of sightings is a product of two distinct processes: the biological process of the species living somewhere, and the observation process of a person being there to record it. The observed pattern is a function of both the species' true distribution and our sampling effort. This simple fact is the central challenge of using presence-only data, a ghost in the machine that can lead our models astray.
If we ignore the observation process, our models can be spectacularly wrong. Imagine mapping records for a rare phantom orchid. If most of our observations come from a single, well-studied national park, our model might conclude that the orchid's ideal habitat is defined by the precise environmental conditions of that park. It doesn't learn about the orchid's niche; it learns about where ecologists like to go hiking. This is sampling bias: the non-random collection of data. The model becomes a map of observer behavior, not species biology.
This bias can be even more subtle and insidious. Consider an ornamental plant native to a mild Australian climate that is now grown in gardens worldwide. Presence records from gardeners might show the plant "surviving" in the arid American Southwest and the cold Northeast. A naive model would interpret this as evidence of an incredibly broad climatic tolerance, suggesting the plant could become a widespread invasive species. But this conclusion is flawed. The plant isn't surviving in a desert; it's surviving in an irrigated, tended garden. It's not surviving a cold winter; it's surviving in a sheltered spot next to a warm house.
This exposes a critical distinction in ecology: the difference between a fundamental niche and a realized niche.
The fundamental niche is the full range of abiotic conditions (like temperature and moisture) under which a species can physiologically survive and reproduce, based on its intrinsic growth rate . Think of this as the species' potential defined in a lab, free from enemies and competitors.
The realized niche is the much smaller subset of those conditions where the species is actually found, constrained by competitors that push it out, predators that eat it, and dispersal barriers like mountains or oceans that it cannot cross.
The garden plant example shows something even more deceptive. The data points don't represent the fundamental or the realized niche; they represent a human-subsidized niche. The model mistakes human intervention for natural resilience, leading to a dangerous overestimation of the species' true capabilities.
How can we, as detectives, correct for the fact that our clues are biased? We can't go back in time and collect them differently, but we can be clever in how we analyze them. The core idea is not to eliminate bias, but to account for it.
Fighting Bias with Bias
One of the most elegant solutions is to use a target-group background. Let's say we're mapping a species of freshwater snail that causes schistosomiasis, and our presence records are all clustered near roads. Instead of comparing the environment at these snail locations to the environment of the entire landscape, we compare it to the locations of all other freshwater species collected by the same programs. This biased background sample acts as a control for the observation process. The logic is powerful: "Given that we are already in a heavily-sampled area near a road, what is special about this specific spot that our target snail likes it here?" By using a background that shares the same sampling bias as our presence data, the bias tends to cancel out, allowing the true ecological signal to shine through.
Modeling the Observer
An alternative approach is to explicitly model the observer's behavior. We can't know the exact path of every naturalist, but we can use effort proxies—measurable variables that are likely correlated with where people look. Covariates like "distance to the nearest road" or "human population density" are often powerful predictors of sampling effort. By including these accessibility variables in our model, we can instruct it to statistically separate the effect of sampling convenience from the effect of genuine environmental suitability. The model learns to answer the question: "How much does this species like high elevations, after accounting for the fact that high elevations are hard to get to and rarely sampled?"
A Brute-Force Fix
A simpler, more direct method is spatial thinning. In our over-sampled national park, we might have hundreds of orchid records clustered together. These points are not independent; they tell us the same thing over and over: "orchids like it here." Thinning reduces this pseudo-replication by enforcing a minimum distance between points, for example, by keeping only one record per square kilometer grid cell. This doesn't fix the problem of un-sampled areas, and it involves throwing away valid data, but it prevents the model's algorithm from being overwhelmed by the high density of points in heavily sampled regions. It's a pragmatic way to give a more equal voice to data from across the species' range.
The type of data we have fundamentally determines the depth of knowledge we can attain. Presence-only data, for all its abundance, sits at the bottom of a hierarchy of certainty.
Level 1: Presence-Only Data With presence-only data, even after applying our clever corrections, we are limited to estimating a Resource Selection Function (RSF). This function tells us about the relative probability of use. We can conclude that a species is twice as likely to select habitat A over habitat B. However, we cannot determine the absolute probability of occurrence. We can't say, "There is a chance the species is present at this location." That absolute scale is hidden from us, hopelessly confounded by the unknown total sampling effort across the entire landscape.
Level 2: Presence-Absence Data A step up is having presence-absence data, where surveyors have recorded not only where they found a species, but also where they looked for it and did not find it. This seems like it should solve everything. But a new problem arises: was an "absence" a true absence, or did the surveyor simply fail to detect a species that was present? The observed outcome is a product: the probability of detecting the species, , is the probability it's actually there, , times the probability you'll find it if it is there, . The data only gives us the product, . We still can't separate the biological reality () from the observation process ().
Level 3: Replicate-Visit Presence-Absence Data The gold standard is detection/non-detection data from replicate visits. Here, surveyors visit the same site multiple times. This is the key that unlocks the puzzle. If you visit a site three times and get detection histories like , you know two things with certainty. First, the species was present at that site (). Second, your detection probability is less than perfect, since you missed it on the second visit. The information from sites with at least one detection allows you to build a model for . Once you have a handle on your own fallibility, you can look at a site with a detection history of and make a much more informed judgment. You can now properly partition the non-detections into "true absences" and "missed presences." This finally allows for the estimation of , the absolute probability of occupancy, which is the ultimate goal for many ecological and conservation questions.
The journey from a simple dot on a map to a calibrated map of probability is a beautiful illustration of the scientific process itself. It reveals how grappling with the inherent limitations and biases in our data forces us to think more deeply, to devise more clever methods, and ultimately, to understand not only the world we are observing, but the very nature of observation itself.
In the previous chapter, we laid bare the machinery for grappling with presence-only data. We saw that a simple collection of points—a list of places where a creature has been seen—is fraught with biases but also pregnant with possibility. Now, we embark on a journey to see what stories these points can tell. It is a journey that will take us from mapping the contemporary world to reconstructing the deep past, from forecasting the futures of species to unraveling the very process of evolution. We will see that the principles for dealing with these humble data points are not just a collection of statistical tricks; they are a unified way of thinking that unlocks profound insights across the entire tapestry of the life sciences.
The most immediate and perhaps most intuitive application of presence-only data is to create a map. Not just a map of where a species has been found, but a map of where it could live. Imagine you are a botanist who has just discovered a new, rare orchid in a few scattered, high-altitude locations. Your first question is, "Where else should we look?" This is the fundamental question of species distribution modeling.
The initial approach is beautifully simple. We take the locations of our orchid and ask, "What do these places have in common?" We might find they all share a similar range of annual temperature, rainfall, and elevation. By painting a map of all the places in the world that share this "climatic signature," we have created our first, preliminary hypothesis of the species' potential distribution. We have used the scattered points of presence to sketch the invisible boundaries of a species' climatic niche—the set of environmental conditions where it can, in principle, survive.
But this first sketch is often crude and can be misleading. Nature, as we know, is far more subtle. And the data we collect are rarely a perfect reflection of it. This brings us to the art and science of refining our picture.
Our maps of species occurrences are almost always biased. We have more records from places that are easy to reach—near roads, towns, and universities—and fewer from remote wilderness. This is the classic "streetlight effect": a drunkard loses his keys in a dark park but searches for them under the lone streetlight, not because that's where he lost them, but because that's where the light is. Similarly, we might mistakenly conclude that a disease-carrying insect prefers to live near clinics, simply because that's where cases are reported and recorded.
How can we possibly disentangle the true environmental preferences of a species from the biases of our own data collection? The solution is remarkably elegant. Instead of comparing the environment at presence locations to the environment of the entire study area, we can compare it to the environment of a "target group." For example, to model one particular species of triatomine bug, we might use the locations of all triatomine bugs in museum collections as our background. The logic is that all these bugs were likely collected with a similar effort and bias. By contrasting our focal species against this biased background, the common sampling bias tends to cancel out, leaving a clearer signal of the species' unique environmental requirements. This clever trick, which is a cornerstone of modern presence-only modeling, is especially vital in our age of "citizen science," where millions of opportunistic records from platforms like iNaturalist or eBird provide immense power, but also immense, spatially structured bias.
Beyond correcting for bias, we can also make our models more biologically intelligent. Instead of just using annual mean temperature, we can ask what truly limits an organism. For an insect, survival through a dry season is paramount. Thus, a variable like "precipitation of the driest quarter" is not just another predictor; it is a direct measure of desiccation risk, a key mechanistic constraint on the insect's life. This shift from purely correlative variables to mechanistic ones is a huge step towards understanding the why behind a species' distribution.
Of course, with a growing number of predictors and increasingly complex statistical tools—from Generalized Linear Models (GLMs) to machine learning behemoths like Boosted Regression Trees and Random Forests—we face a new set of challenges, like handling correlations between predictors and avoiding overfitting. There is no single "best" algorithm; there is a rich toolbox, and the choice of tool depends on whether our goal is pure prediction or deep, interpretable understanding.
Once we have a reliable model linking a species' presence to the environment, we unlock a truly spectacular ability: a form of time travel. The environment is not static. Climates change, continents drift, and sea levels rise and fall. By projecting our models onto maps of past or future environments, we can watch the potential distributions of species shrink, expand, and shift over geological time.
Let's journey back to the Cambrian Period, over 500 million years ago. What was the diversity of life like? The fossil record is our only guide, and it is the ultimate presence-only dataset. A fossil is a definitive proof of presence, but its absence tells us almost nothing. The record is incredibly sparse and biased, punctuated by rare windows of exceptional preservation called Lagerstätten. How, then, can we get a truer picture of ancient biodiversity?
The very same logic we use for modern data applies here. We can think of the probability of finding a fossil as a function of the true ancient distribution and a highly variable "sampling effort"—the chance of preservation. In a low-sampling interval, the probability of detecting a species, even if it was present, might be less than 10%. This means that raw tallies of fossils found in a time bin will drastically underestimate true diversity. Paleontologists have developed methods like "range-through," which assumes a species was present in all the time intervals between its first and last known fossil. But this method, while correcting for non-detection, has its own profound consequences. By filling in gaps, it can artificially depress the estimated rates of evolution—origination and extinction—making the history of life appear more stately and less dynamic than it truly was. Understanding the statistics of presence-only data is therefore not just an ecological tool; it is a lens for reading the history of life on Earth.
We don't need to go back 500 million years to see the power of this approach. Consider a species found on a mainland and a nearby island. Did the island population arise when a few brave individuals colonized it from the mainland (a peripatric, or founder, event)? Or was there once a single, continuous population that was split in two when sea levels rose and created the island (a vicariant event)?
Here, we can perform a breathtaking synthesis of two fields: ecology and genomics. First, we build a distribution model from present-day occurrences and project it onto climate maps of the last Ice Age, when sea levels were lower. This paleo-map might reveal that a land bridge or a corridor of suitable habitat once connected the island and mainland. This is our "stage." Next, we look at the DNA of the organisms, which is the "script" of the evolutionary play. If it was a founder event, the island population's DNA should show the classic signatures of a bottleneck: dramatically reduced genetic diversity () and an excess of rare mutations. If it was a vicariant split, the two populations should have roughly comparable genetic diversity. By fitting a demographic model informed by the ecological "stage" to the genetic "script," we can formally test which story is better supported by the data. The humble presence-only point, when combined with a climate model and a genome, becomes a key piece of evidence in reconstructing the very origins of biodiversity.
If we can project our models into the past, we can also project them into the future. This is one of the most critical applications in our current era of global change. By feeding our models climate scenarios for the year 2070, we can forecast how species' ranges might shift. We can identify future refugia for conservation or, for invasive species, future hotspots of risk. We can even track these shifts in near-real-time. By building separate models from historical (e.g., 1960-1990) and contemporary (1991-2020) presence records, we can quantify how a species' climatic niche may already be changing, potentially tracking climate change by moving to cooler elevations or latitudes.
However, this forecasting ability comes with a profound health warning. Most of these models are correlative. They are based on statistical associations, and they implicitly assume that the factors limiting a species today will be the same ones limiting it in a future, novel climate. This assumption of "niche conservatism" might not always hold.
The scientific frontier is to build mechanistic models. Instead of correlating presence with mean annual temperature, a mechanistic model attempts to simulate the organism's actual physiology. Imagine building a "virtual insect." We would model its life cycle from egg to adult, specifying how its development rate, survival, and fecundity change day-by-day in response to temperature and resource availability. We can then "run" this virtual insect in any location in the world, using daily weather data. If the model shows the insect can successfully complete its life cycle and produce a surplus of offspring (), we predict the population can persist there. This process-based approach, which explicitly models phenology and life-stage-specific requirements, is far more likely to make reliable predictions under novel conditions, as it is based on biological first principles, not just statistical patterns.
We began this story with scattered, opportunistic presence-only points, often disparaged as "dirty" data. We have seen how, with clever analysis, they can paint maps of the living world, reconstruct deep history, and forecast the future. But their final, and perhaps greatest, power lies in their ability to be fused with other data streams.
An ecologist today may have a rich but spatially biased collection of citizen science presence records, alongside a small, rigorous, but spatially limited dataset from a structured monitoring program. In the past, one had to choose which to use. Today, we can have both. Using sophisticated statistical frameworks like state-space models, we can formally integrate these disparate data types. Such a model can use the structured survey data to learn about the probability of detecting a species when it's present, and then use that knowledge to correctly interpret the thousands of presence and non-detection events in the citizen science data. The result is a single, unified estimate of a species' population dynamics that is more robust and has greater coverage than either dataset could provide alone.
From a single dot on a map to a parameter in a unified model of planetary change. This is the intellectual arc of presence-only data. It is a testament to the power of science to extract a universe of understanding from the simplest of observations, turning a whisper of presence into a symphony of ecological and evolutionary insight.