A data set for evaluating the performance of multi-class multi-object video tracking
Date
2017
Authors
Chakraborty, A.
Stamatescu, V.
Wong, S.C.
Wigley, G.
Kearney, D.
Editors
Sadjadi, F.A.
Mahalanobis, A.
Mahalanobis, A.
Advisors
Journal Title
Journal ISSN
Volume Title
Type:
Conference paper
Citation
Proceedings of SPIE, 2017 / Sadjadi, F.A., Mahalanobis, A. (ed./s), vol.10202, iss.102020G, pp.1-9
Statement of Responsibility
Conference Name
Automatic Target Recognition XXVII 2017 (10 Apr 2017 - 17 Apr 2017 : Anaheim, United States)
Abstract
One of the challenges in evaluating multi-object video detection, tracking and classification systems is having publically available data sets with which to compare different systems. However, the measures of performance for tracking and classification are different. Data sets that are suitable for evaluating tracking systems may not be appropriate for classification. Tracking video data sets typically only have ground truth track IDs, while classification video data sets only have ground truth class-label IDs. The former identifies the same object over multiple frames, while the latter identifies the type of object in individual frames. This paper describes an advancement of the ground truth meta-data for the DARPA Neovision2 Tower data set to allow both the evaluation of tracking and classification. The ground truth data sets presented in this paper contain unique object IDs across 5 different classes of object (Car, Bus, Truck, Person, Cyclist) for 24 videos of 871 image frames each. In addition to the object IDs and class labels, the ground truth data also contains the original bounding box coordinates together with new bounding boxes in instances where un-annotated objects were present. The unique IDs are maintained during occlusions between multiple objects or when objects re-enter the field of view. This will provide: A solid foundation for evaluating the performance of multi-object tracking of different types of objects, a straightforward comparison of tracking system performance using the standard Multi Object Tracking (MOT) framework, and classification performance using the Neovision2 metrics. These data have been hosted publically.
School/Discipline
Dissertation Note
Provenance
Description
Data source: Neovision2 Tower dataset, http://ilab.usc.edu/neo2/dataset/tower/training/
Access Status
Rights
Copyright 2017 Society of Photo Optical Instrumentation Engineers (SPIE). One print or electronic copy may be made for personal use only. Systematic reproduction and distribution, duplication of any material in this publication for a fee or for commercial purposes, or modification of the contents of the publication are prohibited.