DroTrack: High-speed drone-based object tracking under uncertainty
Date
2020
Authors
Hamdi, A.
Salim, F.
Kim, D.Y.
Editors
Advisors
Journal Title
Journal ISSN
Volume Title
Type:
Conference paper
Citation
IEEE International Conference on Fuzzy Systems, 2020, vol.2020-July, pp.1-8
Statement of Responsibility
Ali Hamdi, Flora Salim, Du Yong Kim
Conference Name
IEEE International Conference on Fuzzy Systems (FUZZ-IEEE) (19 Jul 2020 - 24 Jul 2020 : Glasgow, UK)
Abstract
We present DroTrack, a high-speed visual singleobject tracking framework for drone-captured video sequences. Most of the existing object tracking methods are designed to tackle well-known challenges, such as occlusion and cluttered backgrounds. The complex motion of drones, i.e., multiple degrees of freedom in three-dimensional space, causes high uncertainty. The uncertainty problem leads to inaccurate location predictions and fuzziness in scale estimations. DroTrack solves such issues by discovering the dependency between object representation and motion geometry. We implement an effective object segmentation based on Fuzzy C Means (FCM). We incorporate the spatial information into the membership function to cluster the most discriminative segments. We then enhance the object segmentation by using a pre-trained Convolution Neural Network (CNN) model. DroTrack also leverages the geometrical angular motion to estimate a reliable object scale. We discuss the experimental results and performance evaluation using two datasets of 51,462 drone-captured frames. The combination of the FCM segmentation and the angular scaling increased DroTrack precision by up to 9% and decreased the centre location error by 162 pixels on average. DroTrack outperforms all the high-speed trackers and achieves comparable results in comparison to deep learning trackers. DroTrack offers high frame rates up to 1000 frame per second (fps) with the best location precision, more than a set of state-of-the-art real-time trackers.
School/Discipline
Dissertation Note
Provenance
Description
Access Status
Rights
©2020 IEEE