Publication:

Improving Post-Training Quantization via Probabilistic Programming

 
cris.virtual.department#PLACEHOLDER_PARENT_METADATA_VALUE#
cris.virtual.department#PLACEHOLDER_PARENT_METADATA_VALUE#
cris.virtual.department#PLACEHOLDER_PARENT_METADATA_VALUE#
cris.virtual.orcid0000-0002-2969-3133
cris.virtual.orcid0000-0003-4456-4353
cris.virtual.orcid0000-0002-1666-5483
cris.virtualsource.department5f457973-5b9f-4593-8a29-1eeb47f32775
cris.virtualsource.departmentf9bc1a0b-4f60-44f9-919f-86e3d0a0b43c
cris.virtualsource.departmentca80dd44-0864-4cd3-a701-870434fb124c
cris.virtualsource.orcid5f457973-5b9f-4593-8a29-1eeb47f32775
cris.virtualsource.orcidf9bc1a0b-4f60-44f9-919f-86e3d0a0b43c
cris.virtualsource.orcidca80dd44-0864-4cd3-a701-870434fb124c
dc.contributor.authorLiu, Kui
dc.contributor.authorGoossens, Bart
dc.contributor.authorDe Schepper, Tom
dc.contributor.authorPhilips, Wilfried
dc.date.accessioned2026-01-08T12:08:19Z
dc.date.available2026-01-08T12:08:19Z
dc.date.issued2025-12
dc.description.abstractPost-training quantization (PTQ) is an effective solution for deploying deep neural networks on edge devices with limited resources. PTQ is especially attractive because it does not require access to the entire original training dataset on the promise of being able to use a much smaller calibration dataset. However, many existing PTQ methods still require a sufficiently large calibration dataset (e.g., more than 1000 images) to achieve satisfactory model accuracy. In this paper, we present a novel post-training quantization method that estimates quantization parameters using a Bayesian Maximum A Posterior (MAP) estimator. By modeling the uncertainty of quantization operations, we formulate the neural network quantization as a Bayesian inference problem. In our method, we first employ probabilistic programming techniques to optimize quantization parameters by maximizing the posterior of quantization step sizes. In addition, we introduce a Minimum Description Length (MDL) prior that favors low quantization bit widths and a validation procedure, which enhances PTQ performance when learning from small calibration datasets. Comprehensive evaluations demonstrate that the proposed method can improve the PTQ performance using a minimal calibration dataset of just 64 images, and achieve nearly state-of-the-art PTQ performance. Furthermore, the proposed method shows strong generalization ability when calibrated on different data sources and tested across diverse data.
dc.identifier10.1109/TCSVT.2025.3588737
dc.identifier.doi10.1109/TCSVT.2025.3588737
dc.identifier.issn1051-8215
dc.identifier.urihttps://imec-publications.be/handle/20.500.12860/58624
dc.identifier.urlhttps://ieeexplore.ieee.org/document/11080033/media#media
dc.language.iso1
dc.provenance.editstepusergreet.vanhoof@imec.be
dc.publisherIEEE
dc.relation.ispartofIEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY
dc.relation.ispartofseriesIEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY
dc.source.beginpage11950
dc.source.endpage11964
dc.source.issue12
dc.source.journalIEEE Transactions on Circuits and Systems for Video Technology
dc.source.numberofpages15
dc.source.volume35
dc.subjectNEURAL-NETWORK QUANTIZATION
dc.subjectQuantization (signal)
dc.subjectTraining
dc.subjectProbabilistic logic
dc.subjectCalibration
dc.subjectNeural networks
dc.subjectOptimization
dc.subjectBayes methods
dc.subjectProgramming
dc.subjectDegradation
dc.subjectAdaptation models
dc.subjectPost-training quantization
dc.subjectBayesian optimization
dc.subjectprobabilistic programming
dc.subjectScience & Technology
dc.subjectTechnology
dc.title

Improving Post-Training Quantization via Probabilistic Programming

dc.typeJournal article
dspace.entity.typePublication
oaire.citation.editionWOS.SCI
oaire.citation.endPage11964
oaire.citation.issue12
oaire.citation.startPage11950
oaire.citation.volume35
person.identifier.ridH-4772-2018
Files

Original bundle

Name:
improving post-training.pdf
Size:
3.33 MB
Format:
Adobe Portable Document Format
Description:
Accepted
Publication available in collections: