Paper published in a book (Scientific congresses, symposiums and conference proceedings)
Q-value functions for decentralized POMDPs
Oliehoek, Frans A.; Vlassis, Nikos
2007In Proc Int. Joint Conf. on Autonomous Agents and Multi-Agent Systems
Peer reviewed
 

Files


Full Text
download.pdf
Publisher postprint (175.6 kB)
Download

All documents in ORBilu are protected by a user license.

Send to



Details



Abstract :
[en] Planning in single-agent models like MDPs and POMDPs can be carried out by resorting to Q-value functions: a (near-) optimal Q-value function is computed in a recursive manner by dynamic programming, and then a policy is extracted from this value function. In this paper we study whether similar Q-value functions can be defined in decentralized POMDP models (Dec-POMDPs), what the cost of computing such value functions is, and how policies can be extracted from such value functions. Using the framework of Bayesian games, we argue that searching for the optimal Q-value function may be as costly as exhaustive policy search. Then we analyze various approximate Q-value functions that allow efficient computation. Finally, we describe a family of algorithms for extracting policies from such Q-value functions.
Disciplines :
Computer science
Identifiers :
UNILU:UL-ARTICLE-2011-710
Author, co-author :
Oliehoek, Frans A.
Vlassis, Nikos ;  University of Luxembourg > Luxembourg Centre for Systems Biomedicine (LCSB)
Language :
English
Title :
Q-value functions for decentralized POMDPs
Publication date :
2007
Event name :
Int. Joint Conf. on Autonomous Agents and Multi-Agent Systems
Event date :
2007
Main work title :
Proc Int. Joint Conf. on Autonomous Agents and Multi-Agent Systems
Pages :
833-840
Peer reviewed :
Peer reviewed
Available on ORBilu :
since 17 November 2013

Statistics


Number of views
48 (2 by Unilu)
Number of downloads
165 (0 by Unilu)

Scopus citations®
 
13
Scopus citations®
without self-citations
8

Bibliography


Similar publications



Contact ORBilu