PriMAT: A robust multi-animal tracking model for primates in the wild
Author:
Vogg RichardORCID, Nuske Matthias, Weis Marissa A.ORCID, Lüddecke Timo, Karakoç Elif, Ahmed ZurnaORCID, Pereira Sofia M., Malaivijitnond Suchinda, Meesawat Suthirote, Wörgötter Florentin, Kappeler Peter M., Gail Alexander, Ostner Julia, Schülke Oliver, Fichtel Claudia, Ecker Alexander S.ORCID
Abstract
AbstractDetection and tracking of animals is an important first step for automated behavioral studies in videos. Nowadays, animal tracking is mostly done using deep learning frameworks based on keypoints, which show remarkable results in lab settings with fixed cameras, backgrounds, and lighting. In contrast, multi-animal tracking in the wild is currently underexplored as it presents numerous challenges such as high background variability, lighting conditions, complex motion, and occlusion.We propose a multi-animal tracking model, PriMAT, for nonhuman primates in the wild. The model learns to detect and track primates and other objects of interest from labeled videos or single images using bounding boxes instead of keypoints. Using bounding boxes significantly facilitates data annotation and allows for more robust results. Our one-stage model is conceptually simple but highly flexible, and we add a classification branch that allows us to train individual identification.To evaluate the performance of our model, we applied it in two case studies with Assamese macaques and redfronted lemurs in the wild. We show that with only a few hundred frames labeled with bounding boxes, we can achieve robust tracking results. Combining these results with the classification branch, our model shows an accuracy of 83% in predicting lemur identities.Our approach presents a promising solution for accurately tracking and identifying animals in the wild using deep learning based bounding box tracking, offering researchers a tool to study animal behavior in their natural habitats. Our code, models, training images, and evaluation video sequences are publicly available (https://github.com/ecker-lab/PriMAT-tracking), facilitating its use for animal behavior analyses and future research in this field.
Publisher
Cold Spring Harbor Laboratory
Reference49 articles.
1. Bain, M. , Nagrani, A. , Schofield, D. , and Zisserman, A. (2019). Count, crop and recognise: Fine-grained recognition in the wild. In Proceedings of the IEEE/CVF International Conference on Computer Vision Workshops. 2. Automated markerless pose estimation in freely moving macaques with openmonkeystudio;Nature communications,2020 3. Evaluating multiple object tracking performance: the clear mot metrics;EURASIP Journal on Image and Video Processing,2008 4. A dataset and application for facial recognition of individual gorillas in zoo environments;arXiv preprint,2020 5. Buchanan, K. , Burt de Perera, T. , Carere, C. , Carter, T. , Hailey, A. , Hubrecht, R. , Jennings, D. , Metcalfe, N. , Pitcher, T. , Peron, F. , et al. (2012). Guidelines for the treatment of animals in behavioural research and teaching.
|
|