Please use this identifier to cite or link to this item:
|Scopus||Web of Science®||Altmetric|
|Title:||Out-of-sample extrapolation of learned manifolds|
|Citation:||IEEE Transactions on Pattern Analysis and Machine Intelligence, 2008; 30(9):1547-1556|
|Publisher:||IEEE Computer Soc|
|Tat-Jun Chin and David Suter|
|Abstract:||We investigate the problem of extrapolating the embedding of a manifold learned from finite samples to novel out-ofsample data. We concentrate on the manifold learning method called Maximum Variance Unfolding (MVU), for which the extrapolation problem is still largely unsolved. Taking the perspective of MVU learning being equivalent to Kernel Principal Component Analysis (KPCA), our problem reduces to extending a kernel matrix generated from an unknown kernel function to novel points. Leveraging on previous developments, we propose a novel solution, which involves approximating the kernel eigenfunction by using Gaussian basis functions. We also show how the width of the Gaussian can be tuned to achieve extrapolation. Experimental results, which demonstrate the effectiveness of the proposed approach, are also included.|
|Keywords:||Manifold learning; out-of-sample extrapolation; Maximum Variance Unfolding|
|Appears in Collections:||Computer Science publications|
Files in This Item:
There are no files associated with this item.
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.