Publication:

Optimistic Reinforcement Learning-Based Skill Insertions for Task and Motion Planning

 
cris.virtual.department#PLACEHOLDER_PARENT_METADATA_VALUE#
cris.virtual.department#PLACEHOLDER_PARENT_METADATA_VALUE#
cris.virtual.department#PLACEHOLDER_PARENT_METADATA_VALUE#
cris.virtual.orcid0000-0002-9063-2751
cris.virtual.orcid0000-0003-4881-9341
cris.virtual.orcid0000-0002-9931-7037
cris.virtualsource.department9575d770-f1f3-41e0-b394-5f6e6e0e81ef
cris.virtualsource.department47530ccc-659e-457a-9b3b-557ce3dd23e7
cris.virtualsource.department94f6cb56-5869-4ef8-b3e9-2a1b5ab44edb
cris.virtualsource.orcid9575d770-f1f3-41e0-b394-5f6e6e0e81ef
cris.virtualsource.orcid47530ccc-659e-457a-9b3b-557ce3dd23e7
cris.virtualsource.orcid94f6cb56-5869-4ef8-b3e9-2a1b5ab44edb
dc.contributor.authorLiu, Gaoyuan
dc.contributor.authorde Winter, Joris
dc.contributor.authorDurodie, Yuri
dc.contributor.authorSteckelmacher, Denis
dc.contributor.authorNowe, Ann
dc.contributor.authorVanderborght, Bram
dc.date.accessioned2026-01-19T14:28:14Z
dc.date.available2026-01-19T14:28:14Z
dc.date.issued2024
dc.description.abstractTask and motion planning (TAMP) for robotics manipulation necessitates long-horizon reasoning involving versatile actions and skills. While deterministic actions can be crafted by sampling or optimizing with certain constraints, planning actions with uncertainty, i.e., probabilistic actions, remains a challenge for TAMP. On the contrary, Reinforcement Learning (RL) excels in acquiring versatile, yet short-horizon, manipulation skills that are robust with uncertainties. In this letter, we design a method that integrates RL skills into TAMP pipelines. Besides the policy, a RL skill is defined with data-driven logical components that enable the skill to be deployed by symbolic planning. A plan refinement sub-routine is designed to further tackle the inevitable effect uncertainties. In the experiments, we compare our method with baseline hierarchical planning from both TAMP and RL fields and illustrate the strength of the method. The results show that by embedding RL skills, we extend the capability of TAMP to domains with probabilistic skills, and improve the planning efficiency compared to the previous methods.
dc.identifier10.1109/LRA.2024.3398402
dc.identifier.doi10.1109/LRA.2024.3398402
dc.identifier.issn2377-3766
dc.identifier.urihttps://imec-publications.be/handle/20.500.12860/58665
dc.language.isoen
dc.provenance.editstepusergreet.vanhoof@imec.be
dc.publisherIEEE
dc.relation.ispartofIEEE ROBOTICS AND AUTOMATION LETTERS
dc.relation.ispartofseriesIEEE ROBOTICS AND AUTOMATION LETTERS
dc.source.beginpage5974
dc.source.endpage5981
dc.source.issue6
dc.source.journalIEEE Robotics and Automation Letters
dc.source.numberofpages8
dc.source.volume9
dc.subjectSAMPLING-BASED METHODS
dc.subjectManipulation planning
dc.subjectreinforcement learning
dc.subjecttask and motion planning
dc.subjectScience & Technology
dc.subjectTechnology
dc.title

Optimistic Reinforcement Learning-Based Skill Insertions for Task and Motion Planning

dc.typeJournal article
dspace.entity.typePublication
oaire.citation.editionWOS.SCI
oaire.citation.endPage5981
oaire.citation.issue6
oaire.citation.startPage5974
oaire.citation.volume9
person.identifier.orcid0000-0002-9063-2751
person.identifier.orcid0000-0002-5818-7539
person.identifier.orcid0000-0003-1521-8494
person.identifier.orcid0000-0001-6346-4564
person.identifier.orcid0000-0003-4881-9341
person.identifier.ridA-1599-2008
person.identifier.rid#PLACEHOLDER_PARENT_METADATA_VALUE#
person.identifier.rid#PLACEHOLDER_PARENT_METADATA_VALUE#
person.identifier.rid#PLACEHOLDER_PARENT_METADATA_VALUE#
person.identifier.rid#PLACEHOLDER_PARENT_METADATA_VALUE#
Files
Publication available in collections: