Open Access. Powered by Scholars. Published by Universities.®

Operations Research, Systems Engineering and Industrial Engineering Commons

Open Access. Powered by Scholars. Published by Universities.®

Series

Artificial Intelligence and Robotics

DEC-POMDP

Publication Year

Articles 1 - 2 of 2

Full-Text Articles in Operations Research, Systems Engineering and Industrial Engineering

Prioritized Shaping Of Models For Solving Dec-Pomdps, Pradeep Reddy Varakantham, William Yeoh, Prasanna Velagapudi, Paul Scerri Jun 2012

Prioritized Shaping Of Models For Solving Dec-Pomdps, Pradeep Reddy Varakantham, William Yeoh, Prasanna Velagapudi, Paul Scerri

Research Collection School Of Information Systems

An interesting class of multi-agent POMDP planning problems can be solved by having agents iteratively solve individual POMDPs, find interactions with other individual plans, shape their transition and reward functions to encourage good interactions and discourage bad ones and then recompute a new plan. D-TREMOR showed that this approach can allow distributed planning for hundreds of agents. However, the quality and speed of the planning process depends on the prioritization scheme used. Lower priority agents shape their models with respect to the models of higher priority agents. In this paper, we introduce a new prioritization scheme that is guaranteed to ...


Distributed Model Shaping For Scaling To Decentralized Pomdps With Hundreds Of Agents, Prasanna Velagapudi, Pradeep Reddy Varakantham, Katia Sycara, Paul Scerri May 2011

Distributed Model Shaping For Scaling To Decentralized Pomdps With Hundreds Of Agents, Prasanna Velagapudi, Pradeep Reddy Varakantham, Katia Sycara, Paul Scerri

Research Collection School Of Information Systems

The use of distributed POMDPs for cooperative teams has been severely limited by the incredibly large joint policy- space that results from combining the policy-spaces of the individual agents. However, much of the computational cost of exploring the entire joint policy space can be avoided by observing that in many domains important interactions between agents occur in a relatively small set of scenarios, previously defined as coordination locales (CLs) [11]. Moreover, even when numerous interactions might occur, given a set of individual policies there are relatively few actual interactions. Exploiting this observation and building on an existing model shaping algorithm ...