Advanced search
1 file | 3.77 MB Add to list

Distributional reinforcement learning-based energy arbitrage strategies in imbalance settlement mechanism

Author
Organization
Project
Abstract
Growth in the penetration of renewable energy sources makes supply more uncertain and leads to an increase in the system imbalance. This trend, together with the single imbalance pricing, opens an opportunity for balance responsible parties (BRPs) to perform energy arbitrage in the imbalance settlement mechanism. To this end, we propose a battery control framework based on distributional reinforcement learning. Our proposed control framework takes a risk-sensitive perspective, allowing BRPs to adjust their risk preferences: we aim to optimize a weighted sum of the arbitrage profit and a risk measure (value-at-risk in this study) while constraining the daily number of cycles for the battery. We assess the performance of our proposed control framework using the Belgian imbalance prices of 2022 and compare two state-of-the-art RL methods, deep Q-learning and soft actor-critic (SAC). Results reveal that the distributional soft actor-critic method outperforms other methods. Moreover, we note that our fully risk-averse agent appropriately learns to hedge against the risk related to the unknown imbalance price by (dis)charging the battery only when the agent is more certain about the price.
Keywords
Battery energy storage systems (BESS), Distributional soft actor-critic (DSAC), Imbalance settlement mechanism, Reinforcement learning (RL), Risk-sensitive energy arbitrage, SERVICES

Downloads

  • (...).pdf
    • full text (Published version)
    • |
    • UGent only
    • |
    • PDF
    • |
    • 3.77 MB

Citation

Please use this url to cite or link to this publication:

MLA
Karimi Madahi, Seyedsoroush, et al. “Distributional Reinforcement Learning-Based Energy Arbitrage Strategies in Imbalance Settlement Mechanism.” JOURNAL OF ENERGY STORAGE, vol. 104, no. Part A, 2024, doi:10.1016/j.est.2024.114377.
APA
Karimi Madahi, S., Claessens, B., & Develder, C. (2024). Distributional reinforcement learning-based energy arbitrage strategies in imbalance settlement mechanism. JOURNAL OF ENERGY STORAGE, 104(Part A). https://doi.org/10.1016/j.est.2024.114377
Chicago author-date
Karimi Madahi, Seyedsoroush, Bert Claessens, and Chris Develder. 2024. “Distributional Reinforcement Learning-Based Energy Arbitrage Strategies in Imbalance Settlement Mechanism.” JOURNAL OF ENERGY STORAGE 104 (Part A). https://doi.org/10.1016/j.est.2024.114377.
Chicago author-date (all authors)
Karimi Madahi, Seyedsoroush, Bert Claessens, and Chris Develder. 2024. “Distributional Reinforcement Learning-Based Energy Arbitrage Strategies in Imbalance Settlement Mechanism.” JOURNAL OF ENERGY STORAGE 104 (Part A). doi:10.1016/j.est.2024.114377.
Vancouver
1.
Karimi Madahi S, Claessens B, Develder C. Distributional reinforcement learning-based energy arbitrage strategies in imbalance settlement mechanism. JOURNAL OF ENERGY STORAGE. 2024;104(Part A).
IEEE
[1]
S. Karimi Madahi, B. Claessens, and C. Develder, “Distributional reinforcement learning-based energy arbitrage strategies in imbalance settlement mechanism,” JOURNAL OF ENERGY STORAGE, vol. 104, no. Part A, 2024.
@article{01JD9RBQR2FPR95NYW9XTW8VKW,
  abstract     = {{Growth in the penetration of renewable energy sources makes supply more uncertain and leads to an increase in the system imbalance. This trend, together with the single imbalance pricing, opens an opportunity for balance responsible parties (BRPs) to perform energy arbitrage in the imbalance settlement mechanism. To this end, we propose a battery control framework based on distributional reinforcement learning. Our proposed control framework takes a risk-sensitive perspective, allowing BRPs to adjust their risk preferences: we aim to optimize a weighted sum of the arbitrage profit and a risk measure (value-at-risk in this study) while constraining the daily number of cycles for the battery. We assess the performance of our proposed control framework using the Belgian imbalance prices of 2022 and compare two state-of-the-art RL methods, deep Q-learning and soft actor-critic (SAC). Results reveal that the distributional soft actor-critic method outperforms other methods. Moreover, we note that our fully risk-averse agent appropriately learns to hedge against the risk related to the unknown imbalance price by (dis)charging the battery only when the agent is more certain about the price.}},
  articleno    = {{114377}},
  author       = {{Karimi Madahi, Seyedsoroush and Claessens, Bert and Develder, Chris}},
  issn         = {{2352-152X}},
  journal      = {{JOURNAL OF ENERGY STORAGE}},
  keywords     = {{Battery energy storage systems (BESS),Distributional soft actor-critic (DSAC),Imbalance settlement mechanism,Reinforcement learning (RL),Risk-sensitive energy arbitrage,SERVICES}},
  language     = {{eng}},
  number       = {{Part A}},
  pages        = {{14}},
  title        = {{Distributional reinforcement learning-based energy arbitrage strategies in imbalance settlement mechanism}},
  url          = {{http://doi.org/10.1016/j.est.2024.114377}},
  volume       = {{104}},
  year         = {{2024}},
}

Altmetric
View in Altmetric
Web of Science
Times cited: