Value-Based Observation Compression for DEC-POMDPs

Alan Carlin and Shlomo Zilberstein. Value-Based Observation Compression for DEC-POMDPs. Proceedings of the Seventh International Conference on Autonomous Agents and Multiagent Systems (AAMAS), 501-508, Estoril, Portugal, 2008.

Abstract

Representing agent policies compactly is essential for improving the scalability of multi-agent planning algorithms. In this paper, we focus on developing a pruning technique that allows us to merge certain observations within agent policies, while minimizing loss of value. This is particularly important for solving finite-horizon decentralized POMDPs, where agent policies are represented as trees, and where the size of policy trees grows exponentially with the number of observations. We introduce a value-based observation compression technique that prunes the least valuable observations while maintaining an error bound on the value lost as a result of pruning. We analyze the characteristics of this pruning strategy and show empirically that it is effective. Thus, we use compact policies to obtain significantly higher values compared with the best existing DEC-POMDP algorithm.

Bibtex entry:

@inproceedings{CZaamas08,
  author	= {Alan Carlin and Shlomo Zilberstein},
  title		= {Value-Based Observation Compression for {DEC-POMDP}s},
  booktitle     = {Proceedings of the Seventh International Conference on Autonomous
                   Agents and Multiagent Systems},
  year		= {2008},
  pages		= {501-508},
  address       = {Estoril, Portugal},
  url		= {http://rbr.cs.umass.edu/shlomo/papers/CZaamas08.html}
}

shlomo@cs.umass.edu
UMass Amherst