Optimal Energy Scheduling of Flexible Industrial Prosumers via Reinforcement Learning

Conference Paper (2023)
Author(s)

Nick van den Bovenkamp (Student TU Delft, Sunrock Investments B.V)

Juan S. Giraldo (TNO)

Edgar Mauricio Salazar (Eindhoven University of Technology)

Pedro P. Vergara Barrios (TU Delft - Intelligent Electrical Power Grids)

Charalambos Konstantinou (King Abdullah University of Science and Technology (KAUST))

P Palensky (TU Delft - Electrical Sustainable Energy)

Research Group
Intelligent Electrical Power Grids
Copyright
© 2023 Nick van den Bovenkamp, Juan S. Giraldo, Edgar Mauricio Salazar Duque, P.P. Vergara Barrios, Charalambos Konstantinou, P. Palensky
DOI related publication
https://doi.org/10.1109/PowerTech55446.2023.10202699
More Info
expand_more
Publication Year
2023
Language
English
Copyright
© 2023 Nick van den Bovenkamp, Juan S. Giraldo, Edgar Mauricio Salazar Duque, P.P. Vergara Barrios, Charalambos Konstantinou, P. Palensky
Research Group
Intelligent Electrical Power Grids
Pages (from-to)
1-6
ISBN (print)
978-1-6654-8779-5
ISBN (electronic)
978-1-6654-8778-8
Reuse Rights

Other than for strictly personal use, it is not permitted to download, forward or distribute the text or part of it, without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license such as Creative Commons.

Abstract

This paper introduces an energy management system (EMS) aiming to minimize electricity operating costs using reinforcement learning (RL) with a linear function approximation. The proposed EMS uses a Q-learning with tile coding (QLTC) algorithm and is compared to a deterministic mixed-integer linear programming (MILP) with perfect forecast information. The comparison is performed using a case study on an industrial manufacturing company in the Netherlands, considering measured electricity consumption, PV generation, and wholesale electricity prices during one week of operation. The results show that the proposed EMS can adjust the prosumer's power consumption considering favorable prices. The electricity costs obtained using the QLTC algorithm are 99% close to those obtained with the MILP model. Furthermore, the results demonstrate that the QLTC model can generalize on previously learned control policies even in the case of missing data and can deploy actions 80% near to the MILP's optimal solution.

Files

Optimal_Energy_Scheduling_of_F... (pdf)
(pdf | 2.9 Mb)
- Embargo expired in 09-02-2024
License info not available