Robust Domain Randomised Reinforcement Learning through Peer-to-Peer Distillation

Chenyang Zhao (University of Edinburgh)*; Timothy Hospedales (Edinburgh University)
PMLR Page

Abstract

In reinforcement learning, domain randomisation is a popular technique for learning general policies that are robust to new environments and domain-shifts at deployment. However, naively aggregating information from randomised domains may lead to high variance in gradient estimation and an unstable learning process. To address this issue, we present a peer-to-peer online distillation strategy for reinforcement learning termed P2PDRL, where multiple learning agents are each assigned to a different environment, and then exchange knowledge through mutual regularisation based on Kullback–Leibler divergence. Our experiments on continuous control tasks show that P2PDRL enables robust learning across a wider randomisation distribution than baselines, and more robust generalisation to new environments at testing.