Bevan, Peggy A., Pantazis, Omiros, Pringle, Holly, Braga Ferreira, Guilherme, Ingram, Daniel J., Madsen, Emily K., Thomas, Liam, Thanet, Dol Raj, Silwal, Thakur, Rayamajhi, Santosh, and others. (2026) Deep learning-based ecological analysis of camera trap images is impacted by training data quality and quantity. Remote Sensing in Ecology and Conservation, . ISSN 2056-3485. (doi:10.1002/rse2.70052) (KAR id:112388)
|
PDF
Publisher pdf
Language: English
This work is licensed under a Creative Commons Attribution 4.0 International License.
|
|
|
Download this file (PDF/1MB) |
Preview |
| Request a format suitable for use with assistive technology e.g. a screenreader | |
|
PDF
Author's Accepted Manuscript
Language: English Restricted to Repository staff only
This work is licensed under a Creative Commons Attribution 4.0 International License.
|
|
|
Contact us about this publication
|
|
| Official URL: https://doi.org/10.1002/rse2.70052 |
|
Abstract
Large image collections generated from camera traps offer valuable insights into species richness, occupancy, and activity patterns, significantly aiding biodiversity monitoring. However, the manual processing of these data sets is time-consuming, hindering analytical processes. To address this, deep neural networks have been widely adopted to automate image labelling, but the impact of classification error on key ecological metrics remains unclear. Here, we analyze data from camera trap collections in an African savannah (82,300 labelled images, 47 species) and an Asian sub-tropical dry forest (40,308 labelled images, 29 species) to compare ecological metrics derived from expert-generated species identifications with those generated by deep-learning classification models. We specifically assess the impact of deep- learning model architecture, the proportion of label noise in the training data, and the size of the training data set on three key ecological metrics: species richness, occupancy, and activity patterns. We found that predictions of species richness derived from deep neural networks closely match those calculated from expert labels and remained resilient to up to 10% noise in the training data set (mis-labelled images) and a 50% reduction in the training data set size. We found that our choice of deep-learning model architecture (ResNet vs. ConvNext-T) or depth (ResNet18, 50, 101) did not impact predicted ecological metrics. In contrast, species-specific metrics were more sensitive; less common and visually similar species were disproportionately affected by a reduction in deep neural network accuracy, with consequences for occupancy and diel activity pattern estimates. To ensure the reliability of their findings, practitioners should prioritize creating large, clean training sets and account for class imbalance across species over exploring numerous deep-learning model architectures.
| Item Type: | Article |
|---|---|
| DOI/Identification number: | 10.1002/rse2.70052 |
| Additional information: | For the purpose of open access, the author has applied a CC BY public copyright licence to any Author Accepted Manuscript version arising from this submission. |
| Uncontrolled keywords: | deep neural networks; computer vision; ecological metrics; occupancy; activity patterns; species richness; camera traps |
| Subjects: | Q Science |
| Institutional Unit: |
Schools > School of Natural Sciences > Conservation Institutes > Durrell Institute of Conservation and Ecology |
| Former Institutional Unit: |
There are no former institutional units.
|
| Funders: | UK Research and Innovation (https://ror.org/001aqnf71) |
| Depositing User: | Daniel Ingram |
| Date Deposited: | 15 Dec 2025 10:44 UTC |
| Last Modified: | 12 Jan 2026 09:15 UTC |
| Resource URI: | https://kar.kent.ac.uk/id/eprint/112388 (The current URI for this page, for reference purposes) |
- Link to SensusAccess
- Export to:
- RefWorks
- EPrints3 XML
- BibTeX
- CSV
- Depositors only (login required):

https://orcid.org/0000-0001-5843-220X
Altmetric
Altmetric