1. Knowing the Past to Predict the Future: Reinforcement Virtual Learning
- Author
-
Zhang, Peng, Huang, Yawen, Hu, Bingzhang, Wang, Shizheng, Duan, Haoran, Moubayed, Noura Al, Zheng, Yefeng, and Long, Yang
- Subjects
FOS: Computer and information sciences ,Computer Science - Machine Learning ,Artificial Intelligence (cs.AI) ,Computer Science - Artificial Intelligence ,FOS: Electrical engineering, electronic engineering, information engineering ,Systems and Control (eess.SY) ,Electrical Engineering and Systems Science - Systems and Control ,Machine Learning (cs.LG) - Abstract
Reinforcement Learning (RL)-based control system has received considerable attention in recent decades. However, in many real-world problems, such as Batch Process Control, the environment is uncertain, which requires expensive interaction to acquire the state and reward values. In this paper, we present a cost-efficient framework, such that the RL model can evolve for itself in a Virtual Space using the predictive models with only historical data. The proposed framework enables a step-by-step RL model to predict the future state and select optimal actions for long-sight decisions. The main focuses are summarized as: 1) how to balance the long-sight and short-sight rewards with an optimal strategy; 2) how to make the virtual model interacting with real environment to converge to a final learning policy. Under the experimental settings of Fed-Batch Process, our method consistently outperforms the existing state-of-the-art methods.
- Published
- 2022
- Full Text
- View/download PDF