Publication:

Quantized Dynamics Models for Hardware-Efficient Control and Planning in Model-Based RL

 
dc.contributor.authorSatya Murthy, Nitish
dc.contributor.authorCatthoor, Francky
dc.contributor.authorVerhelst, Marian
dc.contributor.authorVrancx, Peter
dc.contributor.imecauthorMurthy, Nitish Satya
dc.contributor.imecauthorCatthoor, Francky
dc.contributor.imecauthorVerhelst, Marian
dc.contributor.imecauthorVrancx, Peter
dc.contributor.orcidimecCatthoor, Francky::0000-0002-3599-8515
dc.contributor.orcidimecVerhelst, Marian::0000-0003-3495-9263
dc.contributor.orcidimecVrancx, Peter::0000-0002-9876-3684
dc.date.accessioned2025-04-15T04:20:28Z
dc.date.available2025-04-15T04:20:28Z
dc.date.issued2025
dc.description.abstractReinforcement learning (RL) algorithms can help solve continuous control tasks in robotics. Model-based RL in particular has shown promise for these applications. It can be orders of magnitude more sample-efficient than its model-free counter-part by utilizing Deep Neural Network (DNN) based dynamics models. Furthermore, model-based methods are more robust and more generalizable due to being reward-agnostic. However, the computational complexities involved in planning and control in model-based RL are much higher, causing challenges in real-time deployment on low-resource hardware at the edge. In our work, we focus on reducing the computational footprint of the dynamics models used in model-based RL. To make the algorithm more hardware-efficient, we introduce block floating point data types for DNNs with optimized block performance and different mixed-precision network configurations. The performance impact of these optimizations is assessed across benchmark continuous robotics control tasks. A total memory savings of can be achieved over conventional FP32 networks while reaching comparable rewards in most scenarios.
dc.description.wosFundingTextThis research received funding from the Flemish Government (AI Research Program).
dc.identifier.doi10.1007/978-3-031-74643-7_16
dc.identifier.eisbn978-3-031-74643-7
dc.identifier.isbn978-3-031-74642-0
dc.identifier.issn1865-0929
dc.identifier.urihttps://imec-publications.be/handle/20.500.12860/45534
dc.publisherSPRINGER INTERNATIONAL PUBLISHING AG
dc.source.beginpage196
dc.source.conference8th European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases
dc.source.conferencedate2023-09-18
dc.source.conferencelocationTurin
dc.source.endpage209
dc.source.journalMachine Learning and Principles and Practice of Knowledge Discovery in Databases (ECML PKDD 2023)
dc.source.numberofpages14
dc.title

Quantized Dynamics Models for Hardware-Efficient Control and Planning in Model-Based RL

dc.typeProceedings paper
dspace.entity.typePublication
Files
Publication available in collections: