Learning Constraints From Human Stop-Feedback in Reinforcement Learning

Learning Constraints From Human Stop-Feedback in Reinforcement Learning

Abstract

We investigate an approach for enabling a reinforcement learning agent to learn about dangerous states or constraints from stopfeedback preventing the agent from taking any further, potentially dangerous, actions. Such feedback could be provided by human supervisors overseeing the RL agent’s behavior while carrying out some complex tasks. To enable the RL agent to learn from the supervisor’s feedback, we propose a probabilistic model for approximating how the supervisor’s feedback could have been generated and consider a Bayesian approach for inferring dangerous states. We evaluated our approach using an OpenAI Safety Gym environment and demonstrated that our agent can effectively infer the imposed safety constraints. Furthermore, we conducted a user study to validate our human-inspired feedback model and to obtain insights into the human provision of stop-feedback.

Grafik Top
Authors
  • Poletti, Silvia
  • Testolin, Alberto
  • Tschiatschek, Sebastian
Grafik Top
Shortfacts
Category
Paper in Conference Proceedings or in Workshop Proceedings (Paper)
Event Title
The 22nd International Conference on Autonomous Agents and Multiagent Systems London, 29 May-2 June 2023
Divisions
Data Mining and Machine Learning
Event Location
London, UK
Event Type
Conference
Event Dates
29.05-02.06.2023
Series Name
proceedings of the 22nd International Conference on Autonomous Agents and Multiagent Systems
ISSN/ISBN
978-1-4503-9432-1
Date
29 May 2023
Export
Grafik Top