Publication:
System-Aware Reinforcement Learning for Optimized Implicit Imbalance Participation in Belgium
| dc.contributor.author | Pavirani, Fabio | |
| dc.contributor.author | Karimi Madahi, Seyed Soroush | |
| dc.contributor.author | Claessens, Bert | |
| dc.contributor.author | Develder, Chris | |
| dc.date.accessioned | 2026-03-24T13:34:10Z | |
| dc.date.available | 2026-03-24T13:34:10Z | |
| dc.date.createdwos | 2025-10-29 | |
| dc.date.issued | 2025 | |
| dc.description.abstract | The increasing integration of renewable energy sources into electrical grids has disrupted the balance between production and consumption. To address this challenges, transmission system operators such as the Belgian one have introduced imbalance tariffs that penalize harmful energy deviations. Although the imbalance settlement mechanism allows balance responsible parties to dynamically adjust their energy positions, it also exposes them to significant risks. In fact, Belgian imbalance prices are determined retrospectively at the end of each settlement block, meaning that energy deviations occur under uncertain pricing conditions. Reinforcement learning (RL) offers a promising solution for navigating this uncertainty thanks to its ability to manage stochastic environments and deliver long-term rewards. However, achieving profitable participation in imbalance settlement requires more than just handling price volatility; it also demands a deep understanding of the grid dynamics. This paper examines how enriching an RL agent's observation space with grid-related data can enhance its awareness of system dynamics and improve decision-making. We specifically focus on the agent's performance during unstable periods - i.e., quarters where last-minute deviations in system imbalance heavily influence prices - by introducing a related metric. Using the soft actor-critic algorithm, we control a simulated battery energy storage system participating in the Belgian imbalance settlement, leveraging historical data spanning three years. Our findings indicate that, compared to a system-agnostic RL agent (i.e., an agent that does not have grid-related values in the observation space), the system-aware agents develop more effective policies, particularly during unstable quarters. | |
| dc.description.wosFundingText | This research was partly funded by the Flemish Government under the "Onderzoeksprogramma Artificiele Intelligentie (AI) Vlaanderen" programme and the Horizon Europe project BlueBird (https://bluebird-project.eu/-grant agreement no. 101192452). | |
| dc.identifier.doi | 10.1109/EEM64765.2025.11050139 | |
| dc.identifier.isbn | 979-8-3315-1279-8 | |
| dc.identifier.issn | 2165-4077 | |
| dc.identifier.uri | https://imec-publications.be/handle/20.500.12860/58929 | |
| dc.language.iso | eng | |
| dc.provenance.editstepuser | greet.vanhoof@imec.be | |
| dc.publisher | IEEE | |
| dc.source.conference | 21st International Conference on the European Energy Market (EEM) | |
| dc.source.conferencedate | 2025-05-27 | |
| dc.source.conferencelocation | Lisbon | |
| dc.source.journal | 2025 21ST INTERNATIONAL CONFERENCE ON THE EUROPEAN ENERGY MARKET, EEM | |
| dc.source.numberofpages | 7 | |
| dc.title | System-Aware Reinforcement Learning for Optimized Implicit Imbalance Participation in Belgium | |
| dc.type | Proceedings paper | |
| dspace.entity.type | Publication | |
| imec.internal.crawledAt | 2025-10-22 | |
| imec.internal.source | crawler | |
| Files | ||
| Publication available in collections: |