Publication:
Improving Post-Training Quantization via Probabilistic Programming
| cris.virtual.department | #PLACEHOLDER_PARENT_METADATA_VALUE# | |
| cris.virtual.department | #PLACEHOLDER_PARENT_METADATA_VALUE# | |
| cris.virtual.department | #PLACEHOLDER_PARENT_METADATA_VALUE# | |
| cris.virtual.orcid | 0000-0002-2969-3133 | |
| cris.virtual.orcid | 0000-0003-4456-4353 | |
| cris.virtual.orcid | 0000-0002-1666-5483 | |
| cris.virtualsource.department | 5f457973-5b9f-4593-8a29-1eeb47f32775 | |
| cris.virtualsource.department | f9bc1a0b-4f60-44f9-919f-86e3d0a0b43c | |
| cris.virtualsource.department | ca80dd44-0864-4cd3-a701-870434fb124c | |
| cris.virtualsource.orcid | 5f457973-5b9f-4593-8a29-1eeb47f32775 | |
| cris.virtualsource.orcid | f9bc1a0b-4f60-44f9-919f-86e3d0a0b43c | |
| cris.virtualsource.orcid | ca80dd44-0864-4cd3-a701-870434fb124c | |
| dc.contributor.author | Liu, Kui | |
| dc.contributor.author | Goossens, Bart | |
| dc.contributor.author | De Schepper, Tom | |
| dc.contributor.author | Philips, Wilfried | |
| dc.date.accessioned | 2026-01-08T12:08:19Z | |
| dc.date.available | 2026-01-08T12:08:19Z | |
| dc.date.issued | 2025-12 | |
| dc.description.abstract | Post-training quantization (PTQ) is an effective solution for deploying deep neural networks on edge devices with limited resources. PTQ is especially attractive because it does not require access to the entire original training dataset on the promise of being able to use a much smaller calibration dataset. However, many existing PTQ methods still require a sufficiently large calibration dataset (e.g., more than 1000 images) to achieve satisfactory model accuracy. In this paper, we present a novel post-training quantization method that estimates quantization parameters using a Bayesian Maximum A Posterior (MAP) estimator. By modeling the uncertainty of quantization operations, we formulate the neural network quantization as a Bayesian inference problem. In our method, we first employ probabilistic programming techniques to optimize quantization parameters by maximizing the posterior of quantization step sizes. In addition, we introduce a Minimum Description Length (MDL) prior that favors low quantization bit widths and a validation procedure, which enhances PTQ performance when learning from small calibration datasets. Comprehensive evaluations demonstrate that the proposed method can improve the PTQ performance using a minimal calibration dataset of just 64 images, and achieve nearly state-of-the-art PTQ performance. Furthermore, the proposed method shows strong generalization ability when calibrated on different data sources and tested across diverse data. | |
| dc.identifier | 10.1109/TCSVT.2025.3588737 | |
| dc.identifier.doi | 10.1109/TCSVT.2025.3588737 | |
| dc.identifier.issn | 1051-8215 | |
| dc.identifier.uri | https://imec-publications.be/handle/20.500.12860/58624 | |
| dc.identifier.url | https://ieeexplore.ieee.org/document/11080033/media#media | |
| dc.language.iso | 1 | |
| dc.provenance.editstepuser | greet.vanhoof@imec.be | |
| dc.publisher | IEEE | |
| dc.relation.ispartof | IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY | |
| dc.relation.ispartofseries | IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY | |
| dc.source.beginpage | 11950 | |
| dc.source.endpage | 11964 | |
| dc.source.issue | 12 | |
| dc.source.journal | IEEE Transactions on Circuits and Systems for Video Technology | |
| dc.source.numberofpages | 15 | |
| dc.source.volume | 35 | |
| dc.subject | NEURAL-NETWORK QUANTIZATION | |
| dc.subject | Quantization (signal) | |
| dc.subject | Training | |
| dc.subject | Probabilistic logic | |
| dc.subject | Calibration | |
| dc.subject | Neural networks | |
| dc.subject | Optimization | |
| dc.subject | Bayes methods | |
| dc.subject | Programming | |
| dc.subject | Degradation | |
| dc.subject | Adaptation models | |
| dc.subject | Post-training quantization | |
| dc.subject | Bayesian optimization | |
| dc.subject | probabilistic programming | |
| dc.subject | Science & Technology | |
| dc.subject | Technology | |
| dc.title | Improving Post-Training Quantization via Probabilistic Programming | |
| dc.type | Journal article | |
| dspace.entity.type | Publication | |
| oaire.citation.edition | WOS.SCI | |
| oaire.citation.endPage | 11964 | |
| oaire.citation.issue | 12 | |
| oaire.citation.startPage | 11950 | |
| oaire.citation.volume | 35 | |
| person.identifier.rid | H-4772-2018 | |
| Files | Original bundle
| |
| Publication available in collections: |