Multi-modal self-paced learning for image classification

Date

2018

Authors

Xu, W.
Liu, W.
Huang, X.
Yang, J.
Qiu, S.

Editors

Advisors

Journal Title

Journal ISSN

Volume Title

Type:

Journal article

Citation

Neurocomputing, 2018; 309:134-144

Statement of Responsibility

Wei Xu, Wei Liu, Xiaolin Huang, Jie Yang, Song Qiu

Conference Name

Abstract

Self-paced learning (SPL) is a powerful framework, where samples from easy ones to more complex ones are gradually involved in the learning process. Its superiority is significant when dealing with challenging vision tasks, like natural image classification. However, SPL based image classification can not deal with information from multiple modalities. As images are usually characterized by visual feature descriptors from multiple modalities, only exploiting one of them may lose some complementary information from other modalities. To overcome the above problem, we propose a multi-modal self-paced learning (MSPL) framework for image classification which jointly trains SPL and multi-modal learning into one framework. Specifically, the multi-modal learning process with curriculum information and the curriculum learning process with multi-modal information are iteratively performed until the final mature multi-modal curriculum is learned. As this multi-modal curriculum can grasp the easy to hard knowledge from both the sample level and the modality level, a better model can be learned. Experimental results on four real-world datasets demonstrate the effectiveness of the proposed approach.

School/Discipline

Dissertation Note

Provenance

Description

Access Status

Rights

© 2018 Published by Elsevier B.V.

License

Grant ID

Call number

Persistent link to this record