Skip to main content
Kent Academic Repository

Deep learning-based ecological analysis of camera trap images is impacted by training data quality and quantity

Bevan, Peggy A., Pantazis, Omiros, Pringle, Holly, Braga Ferreira, Guilherme, Ingram, Daniel J., Madsen, Emily K., Thomas, Liam, Thanet, Dol Raj, Silwal, Thakur, Rayamajhi, Santosh, and others. (2026) Deep learning-based ecological analysis of camera trap images is impacted by training data quality and quantity. Remote Sensing in Ecology and Conservation, . ISSN 2056-3485. (doi:10.1002/rse2.70052) (KAR id:112388)

PDF Publisher pdf
Language: English


Download this file
(PDF/1MB)
[thumbnail of Remote Sens Ecol Conserv - 2026 - Bevan - Deep learning‐based ecological analysis of camera trap images is impacted by.pdf]
Preview
Request a format suitable for use with assistive technology e.g. a screenreader
PDF Author's Accepted Manuscript
Language: English

Restricted to Repository staff only

Contact us about this publication
[thumbnail of Deep learning-based ecological analysis of camera trap images is impacted by training data quality and quantity_Ingram_AAM.pdf]
Official URL:
https://doi.org/10.1002/rse2.70052

Abstract

Large image collections generated from camera traps offer valuable insights into species richness, occupancy, and activity patterns, significantly aiding biodiversity monitoring. However, the manual processing of these data sets is time-consuming, hindering analytical processes. To address this, deep neural networks have been widely adopted to automate image labelling, but the impact of classification error on key ecological metrics remains unclear. Here, we analyze data from camera trap collections in an African savannah (82,300 labelled images, 47 species) and an Asian sub-tropical dry forest (40,308 labelled images, 29 species) to compare ecological metrics derived from expert-generated species identifications with those generated by deep-learning classification models. We specifically assess the impact of deep- learning model architecture, the proportion of label noise in the training data, and the size of the training data set on three key ecological metrics: species richness, occupancy, and activity patterns. We found that predictions of species richness derived from deep neural networks closely match those calculated from expert labels and remained resilient to up to 10% noise in the training data set (mis-labelled images) and a 50% reduction in the training data set size. We found that our choice of deep-learning model architecture (ResNet vs. ConvNext-T) or depth (ResNet18, 50, 101) did not impact predicted ecological metrics. In contrast, species-specific metrics were more sensitive; less common and visually similar species were disproportionately affected by a reduction in deep neural network accuracy, with consequences for occupancy and diel activity pattern estimates. To ensure the reliability of their findings, practitioners should prioritize creating large, clean training sets and account for class imbalance across species over exploring numerous deep-learning model architectures.

Item Type: Article
DOI/Identification number: 10.1002/rse2.70052
Additional information: For the purpose of open access, the author has applied a CC BY public copyright licence to any Author Accepted Manuscript version arising from this submission.
Uncontrolled keywords: deep neural networks; computer vision; ecological metrics; occupancy; activity patterns; species richness; camera traps
Subjects: Q Science
Institutional Unit: Schools > School of Natural Sciences > Conservation
Institutes > Durrell Institute of Conservation and Ecology
Former Institutional Unit:
There are no former institutional units.
Funders: UK Research and Innovation (https://ror.org/001aqnf71)
Depositing User: Daniel Ingram
Date Deposited: 15 Dec 2025 10:44 UTC
Last Modified: 12 Jan 2026 09:15 UTC
Resource URI: https://kar.kent.ac.uk/id/eprint/112388 (The current URI for this page, for reference purposes)

University of Kent Author Information

Ingram, Daniel J..

Creator's ORCID: https://orcid.org/0000-0001-5843-220X
CReDIT Contributor Roles: Writing - original draft
  • Depositors only (login required):

Total unique views of this page since July 2020. For more details click on the image.