Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Value-Function Approximations for Partially Observable Markov Decision Processes

About

Partially observable Markov decision processes (POMDPs) provide an elegant mathematical framework for modeling complex decision and planning problems in stochastic domains in which states of the system are observable only indirectly, via a set of imperfect or noisy observations. The modeling advantage of POMDPs, however, comes at a price -- exact methods for solving them are computationally very expensive and thus applicable in practice only to very simple problems. We focus on efficient approximation (heuristic) methods that attempt to alleviate the computational problem and trade off accuracy for speed. We have two objectives here. First, we survey various approximation methods, analyze their properties and relations and provide some new insights into their differences. Second, we present a number of new approximation methods and novel refinements of existing techniques. The theoretical results are supported by experiments on a problem from the agent navigation domain.

M. Hauskrecht• 2011

Related benchmarks

TaskDatasetResultRank
Planning in ACNO-MDPsFrozen Lake 4x4 Default
Total Reward62.42
20
Planning in ACNO-MDPsFrozen Lake 4x4 Hard
Reward Value4.44
20
Planning in ACNO-MDPsFrozen Lake 8x8
Cumulative Reward3.29
20
Showing 3 of 3 rows

Other info

Follow for update