Improving the accessibility and transferability of machine learning algorithms for identification of animals in camera trap images: MLWIC2

Abstract Motion‐activated wildlife cameras (or “camera traps”) are frequently used to remotely and noninvasively observe animals. The vast number of images collected from camera trap projects has prompted some biologists to employ machine learning algorithms to automatically recognize species in the...

Full description

Bibliographic Details
Main Authors: Michael A. Tabak, Mohammad S. Norouzzadeh, David W. Wolfson, Erica J. Newton, Raoul K. Boughton, Jacob S. Ivan, Eric A. Odell, Eric S. Newkirk, Reesa Y. Conrey, Jennifer Stenglein, Fabiola Iannarilli, John Erb, Ryan K. Brook, Amy J. Davis, Jesse Lewis, Daniel P. Walsh, James C. Beasley, Kurt C. VerCauteren, Jeff Clune, Ryan S. Miller
Format: Article
Language:English
Published: Wiley 2020-10-01
Series:Ecology and Evolution
Subjects:
Online Access:https://doi.org/10.1002/ece3.6692
id doaj-b4262503019a4105aecd6166820ad3bb
record_format Article
collection DOAJ
language English
format Article
sources DOAJ
author Michael A. Tabak
Mohammad S. Norouzzadeh
David W. Wolfson
Erica J. Newton
Raoul K. Boughton
Jacob S. Ivan
Eric A. Odell
Eric S. Newkirk
Reesa Y. Conrey
Jennifer Stenglein
Fabiola Iannarilli
John Erb
Ryan K. Brook
Amy J. Davis
Jesse Lewis
Daniel P. Walsh
James C. Beasley
Kurt C. VerCauteren
Jeff Clune
Ryan S. Miller
spellingShingle Michael A. Tabak
Mohammad S. Norouzzadeh
David W. Wolfson
Erica J. Newton
Raoul K. Boughton
Jacob S. Ivan
Eric A. Odell
Eric S. Newkirk
Reesa Y. Conrey
Jennifer Stenglein
Fabiola Iannarilli
John Erb
Ryan K. Brook
Amy J. Davis
Jesse Lewis
Daniel P. Walsh
James C. Beasley
Kurt C. VerCauteren
Jeff Clune
Ryan S. Miller
Improving the accessibility and transferability of machine learning algorithms for identification of animals in camera trap images: MLWIC2
Ecology and Evolution
computer vision
deep convolutional neural networks
image classification
machine learning
motion‐activated camera
R package
author_facet Michael A. Tabak
Mohammad S. Norouzzadeh
David W. Wolfson
Erica J. Newton
Raoul K. Boughton
Jacob S. Ivan
Eric A. Odell
Eric S. Newkirk
Reesa Y. Conrey
Jennifer Stenglein
Fabiola Iannarilli
John Erb
Ryan K. Brook
Amy J. Davis
Jesse Lewis
Daniel P. Walsh
James C. Beasley
Kurt C. VerCauteren
Jeff Clune
Ryan S. Miller
author_sort Michael A. Tabak
title Improving the accessibility and transferability of machine learning algorithms for identification of animals in camera trap images: MLWIC2
title_short Improving the accessibility and transferability of machine learning algorithms for identification of animals in camera trap images: MLWIC2
title_full Improving the accessibility and transferability of machine learning algorithms for identification of animals in camera trap images: MLWIC2
title_fullStr Improving the accessibility and transferability of machine learning algorithms for identification of animals in camera trap images: MLWIC2
title_full_unstemmed Improving the accessibility and transferability of machine learning algorithms for identification of animals in camera trap images: MLWIC2
title_sort improving the accessibility and transferability of machine learning algorithms for identification of animals in camera trap images: mlwic2
publisher Wiley
series Ecology and Evolution
issn 2045-7758
publishDate 2020-10-01
description Abstract Motion‐activated wildlife cameras (or “camera traps”) are frequently used to remotely and noninvasively observe animals. The vast number of images collected from camera trap projects has prompted some biologists to employ machine learning algorithms to automatically recognize species in these images, or at least filter‐out images that do not contain animals. These approaches are often limited by model transferability, as a model trained to recognize species from one location might not work as well for the same species in different locations. Furthermore, these methods often require advanced computational skills, making them inaccessible to many biologists. We used 3 million camera trap images from 18 studies in 10 states across the United States of America to train two deep neural networks, one that recognizes 58 species, the “species model,” and one that determines if an image is empty or if it contains an animal, the “empty‐animal model.” Our species model and empty‐animal model had accuracies of 96.8% and 97.3%, respectively. Furthermore, the models performed well on some out‐of‐sample datasets, as the species model had 91% accuracy on species from Canada (accuracy range 36%–91% across all out‐of‐sample datasets) and the empty‐animal model achieved an accuracy of 91%–94% on out‐of‐sample datasets from different continents. Our software addresses some of the limitations of using machine learning to classify images from camera traps. By including many species from several locations, our species model is potentially applicable to many camera trap studies in North America. We also found that our empty‐animal model can facilitate removal of images without animals globally. We provide the trained models in an R package (MLWIC2: Machine Learning for Wildlife Image Classification in R), which contains Shiny Applications that allow scientists with minimal programming experience to use trained models and train new models in six neural network architectures with varying depths.
topic computer vision
deep convolutional neural networks
image classification
machine learning
motion‐activated camera
R package
url https://doi.org/10.1002/ece3.6692
work_keys_str_mv AT michaelatabak improvingtheaccessibilityandtransferabilityofmachinelearningalgorithmsforidentificationofanimalsincameratrapimagesmlwic2
AT mohammadsnorouzzadeh improvingtheaccessibilityandtransferabilityofmachinelearningalgorithmsforidentificationofanimalsincameratrapimagesmlwic2
AT davidwwolfson improvingtheaccessibilityandtransferabilityofmachinelearningalgorithmsforidentificationofanimalsincameratrapimagesmlwic2
AT ericajnewton improvingtheaccessibilityandtransferabilityofmachinelearningalgorithmsforidentificationofanimalsincameratrapimagesmlwic2
AT raoulkboughton improvingtheaccessibilityandtransferabilityofmachinelearningalgorithmsforidentificationofanimalsincameratrapimagesmlwic2
AT jacobsivan improvingtheaccessibilityandtransferabilityofmachinelearningalgorithmsforidentificationofanimalsincameratrapimagesmlwic2
AT ericaodell improvingtheaccessibilityandtransferabilityofmachinelearningalgorithmsforidentificationofanimalsincameratrapimagesmlwic2
AT ericsnewkirk improvingtheaccessibilityandtransferabilityofmachinelearningalgorithmsforidentificationofanimalsincameratrapimagesmlwic2
AT reesayconrey improvingtheaccessibilityandtransferabilityofmachinelearningalgorithmsforidentificationofanimalsincameratrapimagesmlwic2
AT jenniferstenglein improvingtheaccessibilityandtransferabilityofmachinelearningalgorithmsforidentificationofanimalsincameratrapimagesmlwic2
AT fabiolaiannarilli improvingtheaccessibilityandtransferabilityofmachinelearningalgorithmsforidentificationofanimalsincameratrapimagesmlwic2
AT johnerb improvingtheaccessibilityandtransferabilityofmachinelearningalgorithmsforidentificationofanimalsincameratrapimagesmlwic2
AT ryankbrook improvingtheaccessibilityandtransferabilityofmachinelearningalgorithmsforidentificationofanimalsincameratrapimagesmlwic2
AT amyjdavis improvingtheaccessibilityandtransferabilityofmachinelearningalgorithmsforidentificationofanimalsincameratrapimagesmlwic2
AT jesselewis improvingtheaccessibilityandtransferabilityofmachinelearningalgorithmsforidentificationofanimalsincameratrapimagesmlwic2
AT danielpwalsh improvingtheaccessibilityandtransferabilityofmachinelearningalgorithmsforidentificationofanimalsincameratrapimagesmlwic2
AT jamescbeasley improvingtheaccessibilityandtransferabilityofmachinelearningalgorithmsforidentificationofanimalsincameratrapimagesmlwic2
AT kurtcvercauteren improvingtheaccessibilityandtransferabilityofmachinelearningalgorithmsforidentificationofanimalsincameratrapimagesmlwic2
AT jeffclune improvingtheaccessibilityandtransferabilityofmachinelearningalgorithmsforidentificationofanimalsincameratrapimagesmlwic2
AT ryansmiller improvingtheaccessibilityandtransferabilityofmachinelearningalgorithmsforidentificationofanimalsincameratrapimagesmlwic2
_version_ 1724169560881364992
spelling doaj-b4262503019a4105aecd6166820ad3bb2021-04-02T09:21:12ZengWileyEcology and Evolution2045-77582020-10-011019103741038310.1002/ece3.6692Improving the accessibility and transferability of machine learning algorithms for identification of animals in camera trap images: MLWIC2Michael A. Tabak0Mohammad S. Norouzzadeh1David W. Wolfson2Erica J. Newton3Raoul K. Boughton4Jacob S. Ivan5Eric A. Odell6Eric S. Newkirk7Reesa Y. Conrey8Jennifer Stenglein9Fabiola Iannarilli10John Erb11Ryan K. Brook12Amy J. Davis13Jesse Lewis14Daniel P. Walsh15James C. Beasley16Kurt C. VerCauteren17Jeff Clune18Ryan S. Miller19Quantitative Science Consulting, LLC Laramie WY USAComputer Science Department University of Wyoming Laramie WY USAMinnesota Cooperative Fish and Wildlife Research Unit Department of Fisheries, Wildlife and Conservation Biology University of Minnesota St. Paul MN USAWildlife Research and Monitoring Section Ontario Ministry of Natural Resources and Forestry Peterborough ON CanadaRange Cattle Research and Education Center, Wildlife Ecology and Conservation University of Florida Ona FL USAColorado Parks and Wildlife Fort Collins CO USAColorado Parks and Wildlife Fort Collins CO USAColorado Parks and Wildlife Fort Collins CO USAColorado Parks and Wildlife Fort Collins CO USAWisconsin Department of Natural Resources Madison WI USAConservation Sciences Graduate Program University of Minnesota St. Paul MN USAForest Wildlife Populations and Research Group Minnesota Department of Natural Resources Grand Rapids MN USADepartment of Animal and Poultry Science University of Saskatchewan Saskatoon SK CanadaNational Wildlife Research Center United States Department of Agriculture Fort Collins CO USACollege of Integrative Sciences and Arts Arizona State University Mesa AZ USAUS Geological SurveyNational Wildlife Health Center Madison WI USASavannah River Ecology Laboratory Warnell School of Forestry and Natural Resources University of Georgia Aiken SC USANational Wildlife Research Center United States Department of Agriculture, Animal and Plant Health Inspection Service Fort Collins CO USAOpenAI San Francisco CA USACenter for Epidemiology and Animal Health United States Department of Agriculture Fort Collins CO USAAbstract Motion‐activated wildlife cameras (or “camera traps”) are frequently used to remotely and noninvasively observe animals. The vast number of images collected from camera trap projects has prompted some biologists to employ machine learning algorithms to automatically recognize species in these images, or at least filter‐out images that do not contain animals. These approaches are often limited by model transferability, as a model trained to recognize species from one location might not work as well for the same species in different locations. Furthermore, these methods often require advanced computational skills, making them inaccessible to many biologists. We used 3 million camera trap images from 18 studies in 10 states across the United States of America to train two deep neural networks, one that recognizes 58 species, the “species model,” and one that determines if an image is empty or if it contains an animal, the “empty‐animal model.” Our species model and empty‐animal model had accuracies of 96.8% and 97.3%, respectively. Furthermore, the models performed well on some out‐of‐sample datasets, as the species model had 91% accuracy on species from Canada (accuracy range 36%–91% across all out‐of‐sample datasets) and the empty‐animal model achieved an accuracy of 91%–94% on out‐of‐sample datasets from different continents. Our software addresses some of the limitations of using machine learning to classify images from camera traps. By including many species from several locations, our species model is potentially applicable to many camera trap studies in North America. We also found that our empty‐animal model can facilitate removal of images without animals globally. We provide the trained models in an R package (MLWIC2: Machine Learning for Wildlife Image Classification in R), which contains Shiny Applications that allow scientists with minimal programming experience to use trained models and train new models in six neural network architectures with varying depths.https://doi.org/10.1002/ece3.6692computer visiondeep convolutional neural networksimage classificationmachine learningmotion‐activated cameraR package