Uncertainty quantification is one of the central challenges for machine learning in real-world applications. In reinforcement learning, an agent confronts two kinds of uncertainty, called epistemic uncertainty and aleatoric uncertainty. Disentangling and evaluating these uncertainties simultaneously stands a chance of improving the agent's final performance, accelerating training, and facilitating quality assurance after deployment. In this work, we propose an uncertainty-aware reinforcement learning algorithm for continuous control tasks that extends the Deep Deterministic Policy Gradient algorithm (DDPG). It exploits epistemic uncertainty to accelerate exploration and aleatoric uncertainty to learn a risk-sensitive policy. We conduct numerical experiments showing that our variant of DDPG outperforms vanilla DDPG without uncertainty estimation in benchmark tasks on robotic control and power-grid optimization.
Distributional Actor-Critic Ensemble for Uncertainty-Aware Continuous Control
T. Kanazawa,Haiyan Wang,Chetan Gupta
Published 2022 in IEEE International Joint Conference on Neural Network
ABSTRACT
PUBLICATION RECORD
- Publication year
2022
- Venue
IEEE International Joint Conference on Neural Network
- Publication date
2022-07-18
- Fields of study
Computer Science, Engineering
- Identifiers
- External record
- Source metadata
Semantic Scholar
CITATION MAP
EXTRACTION MAP
CLAIMS
- No claims are published for this paper.
CONCEPTS
- No concepts are published for this paper.
REFERENCES
Showing 1-75 of 75 references · Page 1 of 1
CITED BY
Showing 1-7 of 7 citing papers · Page 1 of 1