Video analysis for the detection of animals using convolutional neural networks and consumer-grade drones
-
Published:2021-06-01
Issue:2
Volume:9
Page:112-127
-
ISSN:2291-3467
-
Container-title:Journal of Unmanned Vehicle Systems
-
language:en
-
Short-container-title:J. Unmanned Veh. Sys.
Author:
Chalmers C.1, Fergus P.1, Curbelo Montanez C. Aday1, Longmore Steven N.2, Wich Serge A.3
Affiliation:
1. School of Computer Science, Liverpool John Moores University, Liverpool L2 2QP, UK. 2. Astrophysics Research Institute, Liverpool John Moores University, Liverpool L3 5RF, UK. 3. School of Biological and Environmental Sciences, Liverpool John Moores University, Liverpool L2 2QP, UK.
Abstract
Determining animal distribution and density is important in conservation. The process is both time-consuming and labour-intensive. Drones have been used to help mitigate human-intensive tasks by covering large geographical areas over a much shorter timescale. In this paper we investigate this idea further using a proof of concept to detect rhinos and cars from drone footage. The proof of concept utilises off-the-shelf technology and consumer-grade drone hardware. The study demonstrates the feasibility of using machine learning (ML) to automate routine conservation tasks, such as animal detection and tracking. The prototype has been developed using a DJI Mavic Pro 2 and tested over a global system for mobile communications (GSM) network. The Faster-RCNN Resnet 101 architecture is used for transfer learning. Inference is performed with a frame sampling technique to address the required trade-off between precision, processing speed, and live video feed synchronisation. Inference models are hosted on a web platform and video streams from the drone (using OcuSync) are transmitted to a real-time messaging protocol (RTMP) server for subsequent classification. During training, the best model achieves a mean average precision (mAP) of 0.83 intersection over union (@IOU) 0.50 and 0.69 @IOU 0.75, respectively. On testing the system in Knowsley Safari our prototype was able to achieve the following: sensitivity (Sen), 0.91 (0.869, 0.94); specificity (Spec), 0.78 (0.74, 0.82); and an accuracy (ACC), 0.84 (0.81, 0.87) when detecting rhinos, and Sen, 1.00 (1.00, 1.00); Spec, 1.00 (1.00, 1.00); and an ACC, 1.00 (1.00, 1.00) when detecting cars.
Publisher
Canadian Science Publishing
Subject
Electrical and Electronic Engineering,Control and Optimization,Computer Science Applications,Aerospace Engineering,Automotive Engineering,Control and Systems Engineering
Reference25 articles.
1. Ba, L.J., and Frey, B. 2013. Adaptive dropout for training deep neural networks. In Advances in neural information processing systems. pp. 3084–3092. 2. Bondi, E., Fang, F., Hamilton, M., Kar, D., Dmello, D., Noronha, V., et al. 2019. Automatic detection of poachers and wildlife with UAVs. In Artificial intelligence and conservation. Edited by F. Fang, M. Tambe, B. Dilkina, and A. Plumptre. Cambridge University Press, Cambridge, UK. pp. 77–100. 3. Buckland, S.T., Anderson, D.R., Burnham, K.P., Laake, J.L., Borchers, D.L., and Thomas, L. 2001. Introduction to distance sampling. Oxford University Press, Oxford, UK. 4. Buckland, S.T., Anderson, D.R., Burnham, K.P., Laake, J.L., Borchers, D.L., and Thomas, L. 2004. Advanced distance sampling: Estimating abundance of biological populations. Oxford University Press, Oxford, UK. 5. Wildlife research and management methods in the 21st century: Where do unmanned aircraft fit in?
Cited by
12 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
|
|