MetaAug: Meta-data Augmentation for Post-training Quantization

dc.contributor.authorPham, C.
dc.contributor.authorHoang, A.D.
dc.contributor.authorNguyen, C.C.
dc.contributor.authorLe, T.
dc.contributor.authorPhung, D.
dc.contributor.authorCarneiro, G.
dc.contributor.authorDo, T.-T.
dc.contributor.conferenceEuropean Conference on Computer Vision (ECCV) (29 Sep 2024 - 4 Oct 2024 : Milan, Italy)
dc.contributor.editorRicci, E.
dc.contributor.editorLeonardis, A.
dc.contributor.editorRoth, S.
dc.contributor.editorRussakovsky, O.
dc.contributor.editorSattler, T.
dc.contributor.editorVarol, G.
dc.date.issued2024
dc.description.abstractPost-Training Quantization (PTQ) has received significant attention because it requires only a small set of calibration data to quantize a full-precision model, which is more practical in real-world applications in which full access to a large training set is not available. However, it often leads to overfitting on the small calibration dataset. Several methods have been proposed to address this issue, yet they still rely on only the calibration set for the quantization and they do not validate the quantized model due to the lack of a validation set. In this work, we propose a novel meta-learning based approach to enhance the performance of post-training quantization. Specifically, to mitigate the overfitting problem, instead of only training the quantized model using the original calibration set without any validation during the learning process as in previous PTQ works, in our approach, we both train and validate the quantized model using two different sets of images. In particular, we propose a meta-learning based approach to jointly optimize a transformation network and a quantized model through bi-level optimization. The transformation network modifies the original calibration data and the modified data will be used as the training set to learn the quantized model with the objective that the quantized model achieves a good performance on the original calibration data. Extensive experiments on the widely used ImageNet dataset with different neural network architectures demonstrate that our approach outperforms the state-of-the-art PTQ methods.
dc.description.statementofresponsibilityCuong Pham, Anh Dung Hoang, Cuong C. Nguyen, Trung Le, Dinh Phung, Gustavo Carneiro, Thanh-Toan Do
dc.identifier.citationLecture Notes in Artificial Intelligence, 2024 / Ricci, E., Leonardis, A., Roth, S., Russakovsky, O., Sattler, T., Varol, G. (ed./s), vol.15085, pp.236-252
dc.identifier.doi10.1007/978-3-031-73383-3_14
dc.identifier.isbn9783031733826
dc.identifier.issn0302-9743
dc.identifier.issn1611-3349
dc.identifier.orcidNguyen, C.C. [0000-0003-2672-6291]
dc.identifier.orcidCarneiro, G. [0000-0002-5571-6220]
dc.identifier.urihttps://hdl.handle.net/2440/144576
dc.language.isoen
dc.publisherSpringer Nature Switzerland
dc.relation.granthttp://purl.org/au-research/grants/arc/DP230101176
dc.relation.ispartofseriesLecture Notes in Computer Science
dc.rights© 2025 The Author(s), under exclusive license to Springer Nature Switzerland AG
dc.source.urihttp://dx.doi.org/10.1007/978-3-031-73383-3_14
dc.subjectPost-Training Quantization (PTQ)
dc.titleMetaAug: Meta-data Augmentation for Post-training Quantization
dc.typeConference paper
pubs.publication-statusPublished

Files

Collections