Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Safe Exploration for Interactive Machine Learning

About

In Interactive Machine Learning (IML), we iteratively make decisions and obtain noisy observations of an unknown function. While IML methods, e.g., Bayesian optimization and active learning, have been successful in applications, on real-world systems they must provably avoid unsafe decisions. To this end, safe IML algorithms must carefully learn about a priori unknown constraints without making unsafe decisions. Existing algorithms for this problem learn about the safety of all decisions to ensure convergence. This is sample-inefficient, as it explores decisions that are not relevant for the original IML objective. In this paper, we introduce a novel framework that renders any existing unsafe IML algorithm safe. Our method works as an add-on that takes suggested decisions as input and exploits regularity assumptions in terms of a Gaussian process prior in order to efficiently learn about their safety. As a result, we only explore the safe set when necessary for the IML problem. We apply our framework to safe Bayesian optimization and to safe exploration in deterministic Markov Decision Processes (MDP), which have been analyzed separately before. Our method outperforms other algorithms empirically.

Matteo Turchetta, Felix Berkenkamp, Andreas Krause• 2019

Related benchmarks

TaskDatasetResultRank
Ground Robot ExplorationGround Robot experiment Continuous Case (test)
Success Rate48.24
4
Safe Object InteractionSafe Object Environment 55 random trials
Success Rate36.36
3
Safe ExplorationGround Robot Environment Discrete Case (80 random trials)
Success Rate35
2
Showing 3 of 3 rows

Other info

Follow for update