A data set for evaluating the performance of multi-class multi-object video tracking

Date

2017

Authors

Chakraborty, A.
Stamatescu, V.
Wong, S.C.
Wigley, G.
Kearney, D.

Editors

Sadjadi, F.A.
Mahalanobis, A.

Advisors

Journal Title

Journal ISSN

Volume Title

Type:

Conference paper

Citation

Proceedings of SPIE, 2017 / Sadjadi, F.A., Mahalanobis, A. (ed./s), vol.10202, iss.102020G, pp.1-9

Statement of Responsibility

Conference Name

Automatic Target Recognition XXVII 2017 (10 Apr 2017 - 17 Apr 2017 : Anaheim, United States)

Abstract

One of the challenges in evaluating multi-object video detection, tracking and classification systems is having publically available data sets with which to compare different systems. However, the measures of performance for tracking and classification are different. Data sets that are suitable for evaluating tracking systems may not be appropriate for classification. Tracking video data sets typically only have ground truth track IDs, while classification video data sets only have ground truth class-label IDs. The former identifies the same object over multiple frames, while the latter identifies the type of object in individual frames. This paper describes an advancement of the ground truth meta-data for the DARPA Neovision2 Tower data set to allow both the evaluation of tracking and classification. The ground truth data sets presented in this paper contain unique object IDs across 5 different classes of object (Car, Bus, Truck, Person, Cyclist) for 24 videos of 871 image frames each. In addition to the object IDs and class labels, the ground truth data also contains the original bounding box coordinates together with new bounding boxes in instances where un-annotated objects were present. The unique IDs are maintained during occlusions between multiple objects or when objects re-enter the field of view. This will provide: A solid foundation for evaluating the performance of multi-object tracking of different types of objects, a straightforward comparison of tracking system performance using the standard Multi Object Tracking (MOT) framework, and classification performance using the Neovision2 metrics. These data have been hosted publically.

School/Discipline

Dissertation Note

Provenance

Description

Data source: Neovision2 Tower dataset, http://ilab.usc.edu/neo2/dataset/tower/training/

Access Status

Rights

Copyright 2017 Society of Photo Optical Instrumentation Engineers (SPIE). One print or electronic copy may be made for personal use only. Systematic reproduction and distribution, duplication of any material in this publication for a fee or for commercial purposes, or modification of the contents of the publication are prohibited.

License

Grant ID

Call number

Persistent link to this record