Publication:

Calibrated Multi-probabilistic Prediction as a Defense Against Adversarial Attacks

 
cris.virtual.department#PLACEHOLDER_PARENT_METADATA_VALUE#
cris.virtual.orcid0000-0002-1666-5483
cris.virtualsource.departmentca80dd44-0864-4cd3-a701-870434fb124c
cris.virtualsource.orcidca80dd44-0864-4cd3-a701-870434fb124c
dc.contributor.authorPeck, Jonathan
dc.contributor.authorGoossens, Bart
dc.contributor.authorSaeys, Yvan
dc.date.accessioned2026-05-11T09:08:09Z
dc.date.available2026-05-11T09:08:09Z
dc.date.createdwos2025-12-10
dc.date.issued2020
dc.description.abstractMachine learning (ML) classifiers—in particular deep neural networks—are surprisingly vulnerable to so-called adversarial examples. These are small modifications of natural inputs which drastically alter the output of the model even though no relevant features appear to have been modified. One explanation that has been offered for this phenomenon is the calibration hypothesis, which states that the probabilistic predictions of typical ML models are miscalibrated. As a result, classifiers can often be very confident in completely erroneous predictions. Based on this idea, we propose the MultIVAP algorithm for defending arbitrary ML models against adversarial examples. Our method is inspired by the inductive Venn-ABERS predictor (IVAP) technique from the field of conformal prediction. The IVAP enjoys the theoretical guarantee that its predictions will be perfectly calibrated, thus addressing the problem of miscalibration. Experimental results on five image classification tasks demonstrate empirically that the MultIVAP has a reasonably small computational overhead and provides significantly higher adversarial robustness without sacrificing accuracy on clean data. This increase in robustness is observed both against defense-oblivious attacks as well as a defense-aware white-box attack specifically designed for the MultIVAP.
dc.description.wosFundingTextWe thank the NVIDIA Corporation for the donation of a Titan Xp GPU with which we were able to carry out our experiments. Jonathan Peck is sponsored by a fellowship of the Research Foundation Flanders (FWO). Yvan Saeys is an ISAC Marylou Ingram scholar.
dc.identifier.doi10.1007/978-3-030-65154-1_6
dc.identifier.isbn978-3-030-65153-4
dc.identifier.issn1865-0929
dc.identifier.urihttps://imec-publications.be/handle/20.500.12860/59411
dc.language.isoeng
dc.provenance.editstepusergreet.vanhoof@imec.be
dc.publisherSPRINGER INTERNATIONAL PUBLISHING AG
dc.source.beginpage85
dc.source.conferenceArtificial Intelligence and Machine Learning , BNAIC 2019, BENELEARN 2019
dc.source.conferencedate2019-11-06
dc.source.conferencelocationBrussels
dc.source.endpage125
dc.source.journalARTIFICIAL INTELLIGENCE AND MACHINE LEARNING, BNAIC 2019, BENELEARN 2019
dc.source.numberofpages41
dc.title

Calibrated Multi-probabilistic Prediction as a Defense Against Adversarial Attacks

dc.typeProceedings paper
dspace.entity.typePublication
imec.internal.crawledAt2026-04-07
imec.internal.sourcecrawler
imec.internal.wosCreatedAt2026-04-07
Files
Publication available in collections: