1. Deep Q-network-based multi-criteria decision-making framework for virtual simulation environment
- Author
-
Hyeonjun Jang, Kyungeun Cho, Yunsick Sung, Shujia Hao, Pradip Kumar Sharma, and Phuong Minh Chu
- Subjects
0209 industrial biotechnology ,Maslow's hierarchy of needs ,Relation (database) ,Computer science ,business.industry ,Deep learning ,Control (management) ,Big data ,02 engineering and technology ,Plan (drawing) ,Expression (mathematics) ,Task (project management) ,020901 industrial engineering & automation ,Artificial Intelligence ,0202 electrical engineering, electronic engineering, information engineering ,Computational Science and Engineering ,020201 artificial intelligence & image processing ,Artificial intelligence ,business ,Software ,Simulation - Abstract
Deep learning improves the realistic expression of virtual simulations specifically to solve multi-criteria decision-making problems, which are generally rely on high-performance artificial intelligence. This study was inspired by the motivation theory and natural life observations. Recently, motivation-based control has been actively studied for realistic expression, but it presents various problems. For instance, it is hard to define the relation among multiple motivations and to select goals based on multiple motivations. Behaviors should generally be practiced to take into account motivations and goals. This paper proposes a deep Q-network (DQN)-based multi-criteria decision-making framework for virtual agents in real time to automatically select goals based on motivations in virtual simulation environments and to plan relevant behaviors to achieve those goals. All motivations are classified according to the five-level Maslow’s hierarchy of needs, and the virtual agents train a double DQN by big social data, select optimal goals depending on motivations, and perform behaviors relying on a predefined hierarchical task networks (HTNs). Compared to the state-of-the-art method, the proposed framework is efficient and reduced the average loss from 0.1239 to 0.0491 and increased accuracy from 63.24 to 80.15%. For behavioral performance using predefined HTNs, the number of methods has increased from 35 in the Q network to 1511 in the proposed framework, and the computation time of 10,000 behavior plans reduced from 0.118 to 0.1079 s.
- Published
- 2020
- Full Text
- View/download PDF