@article{M8AB3E720, title = "Curiosity-Driven TD-MPC with Model-Free Reinfocement Learning Policy", journal = "The Journal of Korean Institute of Communications and Information Sciences", year = "2023", issn = "1226-4717", doi = "10.7840/kics.2023.48.6.712", author = "Chang-Hun Ji, Ju-Bong Kim, Youn-Hee Han", keywords = "Deep Reinforcement Learning, Model-based Reinforcement Learning, Model Predictive Control, Precise Control", abstract = "TD-MPC, which has the highest performance among recent model-based reinforcement learning algorithms, extracts behaviors from model predictive control and DDPG agents in the learning process. However, the DDPG agent does not apply the extracted behavior to the environment, but only applies the behavior extracted from model predicted control to the environment. In this paper, we propose an enhanced TD-MPC that utilizes a dual policy that applies to the environment by considering both the DDPG agent and model predictive control of TD-MPC. In addition, by encouraging exploration based on curiosity, bias in utilization that can occur when choosing an action between dual policies is addressed. It is confirmed that the algorithm proposed in various environments of the DeepMind Control Suite has higher sample efficiency and higher performance than the existing TD-MPC." }