Please use this identifier to cite or link to this item:
|Scopus||Web of Science®||Altmetric|
|Title:||Regularized training of the extreme learning machine using the conjugate gradient method|
|Author:||De Chazal, P.|
|Citation:||2017 International Joint Conference on Neural Networks (IJCNN), 2017 / vol.2017-May, pp.1802-1808|
|Series/Report no.:||IEEE International Joint Conference on Neural Networks (IJCNN)|
|Conference Name:||International Joint Conference on Neural Networks (IJCNN) (14 May 2017 - 19 May 2017 : Anchorage, AK)|
|Philip de Chazal, Mark D. McDonnell|
|Abstract:||We describe a new algorithm providing regularized training of the extreme learning machine (ELM) that uses a modified conjugate gradient (CG) method to determine the network hidden to output weights. The CG method is modified to include a validation set performance calculation at each iteration step. The solution is initialized to zero and during the CG iterations, we monitor the validation set error. When the error begins to rise we terminate the CG algorithm. The operations per iteration is O(P2), where P is the number of output weights, which is significantly faster than the O(P3) operations per iteration required by ridge regression regularization methods. We demonstrate the effectiveness of our method by classifying the MNIST database and achieve an accuracy of 99.2% using an ELM classifier processing the unmodified pixel values.|
|Keywords:||Cholesky; conjugate gradient method; early stopping; Extreme Learning Machine; regularization; QR; SVD|
|Rights:||© 2017 Crown|
|Appears in Collections:||Electrical and Electronic Engineering publications|
Files in This Item:
There are no files associated with this item.
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.