An Analysis of Piecewise-Linear and Convex Value Functions for Active Perception POMDPs

Open Access
Authors
Publication date 2015
Series IAS technical report, IAS-UVA-15-01
Number of pages 16
Publisher Amsterdam: Intelligent Autonomous Systems, University of Amsterdam
Organisations
  • Faculty of Science (FNWI) - Informatics Institute (IVI)
Abstract
In active perception tasks, an agent aims to select actions that reduce its uncertainty about a hidden state. While partially observable Markov decision processes (POMDPs) are a natural model for such problems, reward functions that directly penalize uncertainty in the agent's belief can remove the piecewise-linear and convex (PWLC) property of the value function required by most POMDP planners.
This paper analyses POMDP and POMDP-IR, two frameworks that restore the PWLC property in active perception tasks. We establish the mathematical equivalence of the two frameworks and show that both admit a decomposition of the maximization performed in the Bellman backup, yielding substantial computational savings. We also present an empirical analysis on data from real multi-camera tracking systems that illustrates these savings and analyzes the critical factors in the performance of POMDP planners in such tasks.
Document type Report
Language English
Downloads
pomdpirTechRepo (Final published version)
Permalink to this page
Back