Generalization in Reinforcement Learning with Selective Noise Injection and Information
The ability for policies to generalize to new environments is key to the broadapplication of RL agents. A promising approach to prevent an agent’s policyfrom overfitting to a limited set of training environments is to apply regularizationtechniques originally developed for supervised learning. However, there are starkdifferences between supervised learning and RL. We discuss those differencesand propose modifications to existing regularization techniques in order to betteradapt them to RL. In particular, we focus on regularization techniques relying onthe injection of noise into the learned function, a family that includes some ofthe most widely used approaches such as Dropout and Batch Normalization. Toadapt them to RL, we proposeSelective Noise Injection(SNI), which maintainsthe regularizing effect the injected noise has, while mitigating the adverse effectsit has on the gradient quality. Furthermore, we demonstrate that the InformationBottleneck (IB) is a particularly well suited regularization technique for RL asit is effective in the low-data regime encountered early on in training RL agents.Combining the IB with SNI, we significantly outperform current state of the artresults, including on the recently proposed generalization benchmarkCoinrun.
Top- Igl, Maximilian
- Ciosek, Kamil
- Li, Yingzhen
- Tschiatschek, Sebastian
- Zhang, Cheng
- Devlin, Sam
- Hofmann, Katja
Category |
Paper in Conference Proceedings or in Workshop Proceedings (Paper) |
Event Title |
33ed Conference Neural Information Processing Systems (NeurIPS) |
Divisions |
Data Mining and Machine Learning |
Event Location |
Vancouver, Canada |
Event Type |
Conference |
Event Dates |
08.-14.12.2019 |
Series Name |
Advances in Neural Information Processing Systems 32 (NeurIPS 2019) |
Date |
8 December 2019 |
Export |