Augmenting Cattle Tracking Efficiency Through Monocular Depth Estimation

L.T. Dickson*, C. Davison, C. Michie, E. McRobert, R , Atkinson, I. Andonovic, HJ Ferguson, RJ Dewhurst, R. Briddock, M. Brooking, D Pavlovic, O. Marko, V. Crnojevic, C. Tachtatzis

*Corresponding author for this work

Research output: Contribution to conferencePaperpeer-review

Abstract

We present a method for 3D cattle tracking and
inter-camera pose transformation using depth information from
monocular depth estimation with deep networks. Camerabased animal monitoring offers a minimally invasive and easily
adaptable solution for tracking and welfare monitoring, relying
solely on commercial RGB camera systems. However, environmental factors and inter-animal occlusion often hinder tracking
efficacy and consistency. To address these challenges, we developed a pipeline to extract 3D point cloud data of individual
cows in a straw-bedded calving yard environment, generating
quasi-3D bounding boxes (x, y, z, height, width, θ), where θ is
the polar angle. We then estimate the camera system extrinsic
parameters by minimising the rotation, translation, and scale
discrepancies between the apparent motion of animals across
different frames of reference. This approach demonstrates a
strong agreement between the 3D centroids of tracked animals
in motion. Our work advances the development of algorithmic
occlusion handling and object handover techniques in multicamera systems, particularly pertinent to the high-occlusion,
low-locomotion scenario of animals within barn environments
Original languageEnglish
Publication statusAccepted/In press - 2024
EventCAFE24: IEEE Conference on AgriFood Electronics - Greece
Duration: 26 Sept 202428 Sept 2024

Conference

ConferenceCAFE24
Period26/09/2428/09/24

Fingerprint

Dive into the research topics of 'Augmenting Cattle Tracking Efficiency Through Monocular Depth Estimation'. Together they form a unique fingerprint.

Cite this