Reinforcement learning (RL) can potentially be applied to real-world robot control in complex and uncertain environments. However, it is difficult or even unpractical to design an efficient reward function for various tasks, especially those large and high-dimensional environments. Generative adversarial imitation learning (GAIL) - a general model-free imitation learning method, allows robots to directly learn policies from expert trajectories in large and high-dimensional environments. However, GAIL is still sample inefficient in terms of environmental interaction. In this paper, to solve this problem, we propose a model-based adversarial imitation learning from demonstrations and human reward (MAILDH), a novel model-based interactive imitation framework combining the advantages of GAIL, interactive RL and model-based RL. We tested our method in eight physics-based discrete and continuous control tasks for RL. Our results show that MAILDH can greatly improve the sample efficiency and robustness compared to the original GAIL.
Model-based Adversarial Imitation Learning from Demonstrations and Human Reward
Jie Huang,Jiangshan Hao,Rongshun Juan,Randy Gomez,Keisuke Nakarnura,Guangliang Li
Published 2023 in IEEE/RJS International Conference on Intelligent RObots and Systems
ABSTRACT
PUBLICATION RECORD
- Publication year
2023
- Venue
IEEE/RJS International Conference on Intelligent RObots and Systems
- Publication date
2023-10-01
- Fields of study
Computer Science, Engineering
- Identifiers
- External record
- Source metadata
Semantic Scholar
CITATION MAP
EXTRACTION MAP
CLAIMS
- No claims are published for this paper.
CONCEPTS
- No concepts are published for this paper.
REFERENCES
Showing 1-49 of 49 references · Page 1 of 1
CITED BY
Showing 1-3 of 3 citing papers · Page 1 of 1