Back to Search Start Over

Control of superheat of organic Rankine cycle under transient heat source based on deep reinforcement learning.

Authors :
Wang, Xuan
Wang, Rui
Jin, Ming
Shu, Gequn
Tian, Hua
Pan, Jiaying
Source :
Applied Energy. Nov2020, Vol. 278, pN.PAG-N.PAG. 1p.
Publication Year :
2020

Abstract

• DRL-based control methods are proposed for ORC for engine waste heat recovery. • The DRL control for ORC superheat performs considerably better than PID control. • The DRL-based PID control is more robust than the DRL control. • The switching DRL control performs well and exhibits sufficient robustness. • Useful reference and motivation for application of DRL to thermodynamic systems. The organic Rankine cycle (ORC) is a promising technology for engine waste heat recovery. During real-world operation, the engine working condition varies frequently to satisfy the power demand; thus, the transient nature of engine waste heat presents significant control challenges for the ORC. To control the superheat of the ORC precisely under a transient heat source, several optimal control methods have been used such as model predictive control and dynamic programing. However, most of them depend strongly on the accurate prediction of future disturbances. Deep reinforcement learning (DRL) is an artificial-intelligence algorithm that can overcome the aforementioned disadvantage, but the potential of DRL in control of thermodynamic systems has not yet been investigated. Thus, this paper proposes two DRL-based control methods for controlling the superheat of ORC under a transient heat source. One directly uses the DRL agent to learn the control strategy (DRL control), and the other uses the DRL agent to optimize the parameters of the proportional–integral–derivative (PID) controller (DRL-based PID control). Additionally, a switching mechanism between different DRL controllers is proposed for improving the training efficiency and enlarging the operation range of the controller. The results of this study indicate that the DRL agent can satisfactorily perform the control task and optimize the traditional controller under the trained and untrained transient heat source. Specifically, the DRL control can track the reference superheat with an average error of only 0.19 K, whereas that of the traditional PID control is 2.16 K. Furthermore, the proposed switching DRL control exhibits excellent tracking performance with an average error of only 0.21 K and robustness over a wide range of operation conditions. The successful application of DRL demonstrates its considerable potential for the control of thermodynamic systems, providing a useful reference and motivation for the application to other thermodynamic systems. [ABSTRACT FROM AUTHOR]

Details

Language :
English
ISSN :
03062619
Volume :
278
Database :
Academic Search Index
Journal :
Applied Energy
Publication Type :
Academic Journal
Accession number :
147202896
Full Text :
https://doi.org/10.1016/j.apenergy.2020.115637