Journal of Frontiers of Computer Science and Technology ›› 2020, Vol. 14 ›› Issue (6): 918-927.DOI: 10.3778/j.issn.1673-9418.1912040

Previous Articles     Next Articles

Review of Model-Based Reinforcement Learning

ZHAO Tingting, KONG Le, HAN Yajie, REN Dehua, CHEN Yarui   

  1. College of Artificial Intelligence, Tianjin University of Science and Technology, Tianjin 300467, China
  • Online:2020-06-01 Published:2020-06-04



  1. 天津科技大学 人工智能学院,天津 300467


Deep reinforcement learning (DRL) as an important learning paradigm in the field of machine learning, has received increasing attentions after AlphaGo defeats the human. DRL interacts with the environment by trials and errors, and obtains the optimal policy by maximizing the cumulative reward. Reinforcement learning can be divided into two categories: model-free reinforcement learning and model-based reinforcement learning. The tra-ining process of model-free reinforcement learning needs a large number of samples. It is difficult for model-free reinforcement learning to get good performance when the sampling budget is limited, and a large number of samples cannot be collected. However, model-based reinforcement learning can reduce the real sample demand and improve the data efficiency through making full use of the environment model. This paper focuses on the field of model-based reinforcement learning, introduces its research status, investigates its classical algorithms, and discusses?future development trend and application prospect.

Key words: deep reinforcement learning (DRL), model-based reinforcement learning, state transition model, sample efficiency



关键词: 深度强化学习(DRL), 模型化强化学习, 状态转移模型, 样本利用率