Integrated forecasting and deep reinforcement learning for price-based self-scheduling of PV-BESS: Utility-scale evidence in Chile

基于价格的光伏储能系统自调度集成预测和深度强化学习:智利公用事业规模实例

阅读:1

Abstract

Deep Reinforcement Learning (DRL) shows good performance for optimizing battery energy storage systems (BESS) coordinated operations with photovoltaic plants (PV), yet most studies rely on simulations. Bridging the gap to practical application requires validation using real-world operational data. This paper provides such empirical evidence by developing and rigorously evaluating an integrated forecast-and-control framework on three distinct utility-scale PV-BESS assets in Chile. The framework couples a Sequence-to-Sequence (Seq2Seq) LSTM for point forecaster embedded in a probabilistic scenario-generation pipeline of PV generation with nodal prices with DRL agents (Proximal Policy Optimization - PPO and Soft Actor-Critic - SAC) trained on 1,000 generated scenarios per site. Using two years (2022-2023) of operational plant data, meteorology, and market prices, we benchmark DRL policies against theoretical limits (Oracle), a deterministic predict-then-optimize baseline, a scenario-based model predictive control (MPC), and a random Dummy policy over 14-day horizons using a 900/100 train-test split. The Seq2Seq forecaster improves accuracy (e.g., 34.5% reduction in RMSE for prices vs. SARIMAX). We find that the DRL agents consistently outperform the predict-then-optimize baseline, achieving mean 14-day profits near USD 55k, and exhibiting robust, adaptive contracyclical behavior without excessive cycling. Our study provides a reproducible blueprint and empirical validation for data-driven BESS control, demonstrating its practical viability and economic benefits in real-world operating conditions.

特别声明

1、本页面内容包含部分的内容是基于公开信息的合理引用;引用内容仅为补充信息,不代表本站立场。

2、若认为本页面引用内容涉及侵权,请及时与本站联系,我们将第一时间处理。

3、其他媒体/个人如需使用本页面原创内容,需注明“来源:[生知库]”并获得授权;使用引用内容的,需自行联系原作者获得许可。

4、投稿及合作请联系:info@biocloudy.com。