Publication:
The increasing integration of renewable energy sources into electrical grids has disrupted the balance between production and consumption. To address this challenges, transmission system operators such as the Belgian one have introduced imbalance tariffs that penalize harmful energy deviations. Although the imbalance settlement mechanism allows balance responsible parties to dynamically adjust their energy positions, it also exposes them to significant risks. In fact, Belgian imbalance prices are determined retrospectively at the end of each settlement block, meaning that energy deviations occur under uncertain pricing conditions. Reinforcement learning (RL) offers a promising solution for navigating this uncertainty thanks to its ability to manage stochastic environments and deliver long-term rewards. However, achieving profitable participation in imbalance settlement requires more than just handling price volatility; it also demands a deep understanding of the grid dynamics. This paper examines how enriching an RL agent's observation space with grid-related data can enhance its awareness of system dynamics and improve decision-making. We specifically focus on the agent's performance during unstable periods