Back to Search Start Over

Novel Data-Driven decentralized coordination model for electric vehicle aggregator and energy hub entities in multi-energy system using an improved multi-agent DRL approach.

Authors :
Zhang, Bin
Hu, Weihao
Cao, Di
Ghias, Amer M.Y.M.
Chen, Zhe
Source :
Applied Energy. Jun2023, Vol. 339, pN.PAG-N.PAG. 1p.
Publication Year :
2023

Abstract

• Model-free decentralized coordination control of electric vehicle aggregator and energy hub entities is proposed. • Novel multi-agent DRL framework for continuous control is utilized. • LSTM and safety networks are introduced to stabilize the training process. • Developed algorithm exhibits effective and fast control against uncertainties. • Developed method outperforms 3 advanced benchmark algorithms. Energy hub (EH) is an independent entity that benefits to the efficiency, flexibility, and reliability of integrated energy systems (IESs). On the other hand, the rapid emerging of electric vehicles (EVs) drives the EV aggregator (EVAGG) as another independent entity to facilitate the electricity exchange with the grid. However, due to privacy consideration for different owners, it is challenging to investigate the optimal coordinated strategies for such interconnected entities only by exchanging the information of electrical energy. Besides, the existence of parameter uncertainties (load demands, EVs' charging behaviors, wind power and photovoltaic generation), continuous decision space, dynamic energy flows, and non-convex multi-objective function is difficult to solve. To this end, this paper proposes a novel model-free multi-agent deep reinforcement learning (MADRL) -based decentralized coordination model to minimize the energy costs of EH entities and maximize profits of EVAGGs. First, a long short-term memory (LSTM) module is used to capture the future trend of uncertainties. Then, the coordination problem is formulated as Markov games and solved by the attention enabled MADRL algorithm, where the EH or EVAGG entity is modeled as an adaptive agent. An attention mechanism makes each agent only focus on state information related to the reward. The proposed MADRL adopts the forms of offline centralized training to learn the optimal coordinated control strategy, and decentralized execution to enable agents' online decisions to only require local measurements. A safety network is employed to cope with equality constraints (demand–supply balance). Simulation results illustrate that the proposed method achieves similar results compared to the traditional model-based method with perfect knowledge of system models, and the computation performance is at least two orders of magnitudes shorter than the traditional method. The testing results of the proposed method are better than those of the Concurrent and other MADRL method, with 10.79%/3.06% lower energy cost and 17.11%/6.82% higher profits of aggregator. Besides, the electric equality constraint of the proposed method is only 0.25 MW averaged per day, which is a small and acceptable violation. [ABSTRACT FROM AUTHOR]

Details

Language :
English
ISSN :
03062619
Volume :
339
Database :
Academic Search Index
Journal :
Applied Energy
Publication Type :
Academic Journal
Accession number :
163187987
Full Text :
https://doi.org/10.1016/j.apenergy.2023.120902