The IKEA ASM Dataset: Understanding people assembling furniture through actions, objects and pose
Date
2021
Authors
Ben-Shabat, Y.
Yu, X.
Saleh, F.
Campbell, D.
Rodriguez Opazo, C.
Li, H.
Gould, S.
Editors
Advisors
Journal Title
Journal ISSN
Volume Title
Type:
Conference paper
Citation
Proceedings of the IEEE Winter Conference on Applications of Computer Vision (WACV 2021), 2021, pp.846-858
Statement of Responsibility
Yizhak Ben-Shabat, Xin Yu, Fatemeh Saleh, Dylan Campbell, Cristian Rodriguez-Opazo, Hongdong Li, Stephen Gould
Conference Name
IEEE Winter Conference on Applications of Computer Vision (WACV) (3 Jan 2021 - 9 Jan 2021 : virtual online)
Abstract
The availability of a large labeled dataset is a key requirement for applying deep learning methods to solve various computer vision tasks. In the context of understanding human activities, existing public datasets, while large in size, are often limited to a single RGB camera and provide only per-frame or per-clip action annotations. To enable richer analysis and understanding of human activities, we introduce IKEA ASM—a three million frame, multi-view, furniture assembly video dataset that includes depth, atomic actions, object segmentation, and human poses. Additionally, we benchmark prominent methods for video action recognition, object segmentation and human pose estimation tasks on this challenging dataset. The dataset enables the development of holistic methods, which integrate multimodal and multi-view data to better perform on these tasks.
School/Discipline
Dissertation Note
Provenance
Description
Access Status
Rights
©2021 IEEE