Mapping Species From Crowdsourced Data Using Machine Learning

Using machine learning to generate geographical range predictions for tens of thousands of species with support from the Climate Change AI Innovation Grants Program.

Innovation Grants Ecosystems & Biodiversity
Visualization of features learned by the species mapping approach.

The users of community science platforms such as iNaturalist (www.inaturalist.org) generate millions of photographic observations each month documenting where different plant and animal species can be found. In the last few years, advances in AI in the form of automated image classifiers allow non-experts to identify the different species that are present in these images. However, automatic species identification in images remains a challenging problem, as community science platforms can potentially contain images from hundreds of thousands of different species. One of the major sources of difficulty is that there can be multiple species that look visually similar (e.g. crows versus ravens), and thus remain very challenging for AI systems, and humans, to visually disambiguate.

Knowing where a species of interest is likely to be observed (i.e. its geographical range) is a valuable piece of information when trying to determine what species might be present in an image. Inspired by this simple observation, in recent work, we have developed efficient machine learning models that can take the location where an image was captured as input to determine which species are likely to be present at that geographical location. These models are trained on the locations of observations from iNaturalist, though there are many other platforms that it can be applied to.

Predictions about which species occur at a specific location can be combined with the predictions from automated image classifiers to enhance the accuracy of the final species estimate. We have found that this approach improves species classification performance in images from 75% to 82% when evaluated on a dataset that contains images from 40,000 different species from around the world. Importantly, this improvement does not necessitate any changes in the underlying image classifier being used.

Image of Toad with Maps
Does the image in the top left, taken in Central Europe, contain a European Toad or a Spiny Toad? By looking at the past history of such observations on iNaturalist, shown in red below, we can see that it is very unlikely for Spiny Toads to be found outside of Portugal, Spain, or France.

Motivated by this success, we next wanted to understand how effective these models trained on crowdsourced location data are at predicting the ranges of species on a global scale. Unfortunately, unlike in the image classification example above where we could easily measure the improvement in the final image classification performance, assessing the accuracy of a predicted range for a species is a much more challenging problem. This is because, with some exceptions, there are no established benchmarks for this task.

To address this problem, we developed a suite of spatial prediction tasks for evaluating the performance of machine learning-based species range estimation models. Our new benchmark will enable researchers to evaluate different, and complementary, spatial prediction tasks, from estimating species’ ranges to quantifying how well these models help resolve mistakes in image classifiers. Despite being trained on spatially biased and incomplete data, we were able to show that it is possible to obtain performance that is within 75-80% agreement of expert-derived ranges across a wide set of species. These are promising initial findings, but also point to the fact that there is more work to be done to further improve performance on these spatial prediction tasks.

Image of four prediction maps
Example predictions from one of our machine learning-based species range estimation models. Here we illustrate the results for four different species, where brighter colors indicate locations where the model is more confident that the species occurs. For reference, we add the map of current iNaturalist observations in the bottom right of each.

Main findings:

Gif of specifes ranges
Our machine learning approach jointly encodes the ranges of thousands of different species into one compact model. On the left we visualize the learned representations for 47,000 different species in two dimensions, and on the right we can see the corresponding estimated ranges for a subset of them denoted by the red dot on the left.

Going forward there are many interesting open questions that require additional investigation. For example, species observations from community science platforms can exhibit large geographical and temporal biases as some locations can be difficult for community scientists to reach. In addition, there are global biases as some of the most biodiverse locations on the planet are not yet well represented on these platforms. Some species are also much easier to photograph and identify than others; thus, detectability and data imbalance needs to be explicitly factored into our models. Care needs to be taken so that potential mistakes made by the range estimation models do not adversely impact the image classifiers and result in inaccurate identifications being made. These open questions mean that great caution should be exercised if attempting to make any conservation decisions based on the outputs of these current models.

In the future, we can expect to see more species observation data being uploaded to platforms such as iNaturalist. As the global reach of these online communities grows, we will also potentially obtain more data from currently underreported, but highly biodiverse, regions. Addressing the open methodological issues highlighted above will lead to a new set of methods that will be able to use this data to reliably estimate the ranges of previously mapped species. The resulting improved understanding of how different species are spatially distributed around the world will be an important piece of information in our quest to better conserve them.

More information about this work can be found in our recent paper titled “Spatial Implicit Neural Representations for Global-Scale Species Mapping” (https://arxiv.org/abs/2306.02564). The paper includes a detailed description of the models and our benchmark evaluation datasets and also includes a detailed discussion of the results. The work was presented at the International Conference on Machine Learning in July 2023.

Team members:

Image of team
This post represents the views of its authors, and does not necessarily represent the views of Climate Change AI.