Back to Search Start Over

Insights and approaches using deep learning to classify wildlife.

Authors :
Miao, Zhongqi
Gaynor, Kaitlyn M.
Wang, Jiayun
Liu, Ziwei
Muellerklein, Oliver
Norouzzadeh, Mohammad Sadegh
McInturff, Alex
Bowie, Rauri C. K.
Nathan, Ran
Yu, Stella X.
Getz, Wayne M.
Source :
Scientific Reports; 5/31/2019, Vol. 9 Issue 1, pN.PAG-N.PAG, 1p
Publication Year :
2019

Abstract

The implementation of intelligent software to identify and classify objects and individuals in visual fields is a technology of growing importance to operatives in many fields, including wildlife conservation and management. To non-experts, the methods can be abstruse and the results mystifying. Here, in the context of applying cutting edge methods to classify wildlife species from camera-trap data, we shed light on the methods themselves and types of features these methods extract to make efficient identifications and reliable classifications. The current state of the art is to employ convolutional neural networks (CNN) encoded within deep-learning algorithms. We outline these methods and present results obtained in training a CNN to classify 20 African wildlife species with an overall accuracy of 87.5% from a dataset containing 111,467 images. We demonstrate the application of a gradient-weighted class-activation-mapping (Grad-CAM) procedure to extract the most salient pixels in the final convolution layer. We show that these pixels highlight features in particular images that in some cases are similar to those used to train humans to identify these species. Further, we used mutual information methods to identify the neurons in the final convolution layer that consistently respond most strongly across a set of images of one particular species. We then interpret the features in the image where the strongest responses occur, and present dataset biases that were revealed by these extracted features. We also used hierarchical clustering of feature vectors (i.e., the state of the final fully-connected layer in the CNN) associated with each image to produce a visual similarity dendrogram of identified species. Finally, we evaluated the relative unfamiliarity of images that were not part of the training set when these images were one of the 20 species "known" to our CNN in contrast to images of the species that were "unknown" to our CNN. [ABSTRACT FROM AUTHOR]

Details

Language :
English
ISSN :
20452322
Volume :
9
Issue :
1
Database :
Complementary Index
Journal :
Scientific Reports
Publication Type :
Academic Journal
Accession number :
136768595
Full Text :
https://doi.org/10.1038/s41598-019-44565-w