Doshi, Prashant
On Markov Games Played by Bayesian and Boundedly-Rational Players
Chandrasekaran, Muthukumaran (University of Georgia) | Chen, Yingke (Sichuan University) | Doshi, Prashant (University of Georgia)
We present a new game-theoretic framework in which Bayesian players with bounded rationality engage in a Markov game and each has private but incomplete information regarding other players' types. Instead of utilizing Harsanyi's abstract types and a common prior, we construct intentional player types whose structure is explicit and induces a {\em finite-level} belief hierarchy. We characterize an equilibrium in this game and establish the conditions for existence of the equilibrium. The computation of finding such equilibria is formalized as a constraint satisfaction problem and its effectiveness is demonstrated on two cooperative domains.
Dynamic Sum Product Networks for Tractable Inference on Sequence Data (Extended Version)
Melibari, Mazen, Poupart, Pascal, Doshi, Prashant, Trimponias, George
Sum-Product Networks (SPN) have recently emerged as a new class of tractable probabilistic graphical models. Unlike Bayesian networks and Markov networks where inference may be exponential in the size of the network, inference in SPNs is in time linear in the size of the network. Since SPNs represent distributions over a fixed set of variables only, we propose dynamic sum product networks (DSPNs) as a generalization of SPNs for sequence data of varying length. A DSPN consists of a template network that is repeated as many times as needed to model data sequences of any length. We present a local search technique to learn the structure of the template network. In contrast to dynamic Bayesian networks for which inference is generally exponential in the number of variables per time slice, DSPNs inherit the linear inference complexity of SPNs. We demonstrate the advantages of DSPNs over DBNs and other models on several datasets of sequence data.
Decision Sum-Product-Max Networks
Melibari, Mazen (University of Waterloo) | Poupart, Pascal (University of Waterloo) | Doshi, Prashant (University of Georgia)
Sum-Product Networks (SPNs) were recently proposed as a new class of probabilistic graphical models that guarantee tractable inference, even on models with high-treewidth. In this paper, we propose a new extension to SPNs, called Decision Sum-Product-Max Networks (Decision-SPMNs), that makes SPNs suitable for discrete multi-stage decision problems. We present an algorithm that solves Decision-SPMNs in a time that is linear in the size of the network. We also present algorithms to learn the parameters of the network from data.
Bayesian Markov Games with Explicit Finite-Level Types
Chandrasekaran, Muthukumaran (University of Georgia) | Chen, Yingke (University of Georgia) | Doshi, Prashant (University of Georgia)
In impromptu or ad hoc settings, participating players are precluded from precoordination. Subsequently, each player's own model is private and includes some uncertainty about the others' types or behaviors. Harsanyi's formulation of a Bayesian game lays emphasis on this uncertainty while the players each play exactly one turn. We propose a new game-theoretic framework where Bayesian players engage in a Markov game and each has private but imperfect information regarding other players' types. Consequently, we construct player types whose structure is explicit and includes a finite level belief hierarchy instead of utilizing Harsanyi's abstract types and a common prior distribution. We formalize this new framework and demonstrate its effectiveness on two standard ad hoc teamwork domains involving two or more ad hoc players.
Bayesian Markov Games with Explicit Finite-Level Types
Chandrasekaran, Muthukumaran (University of Georgia) | Chen, Yingke (University of Georgia) | Doshi, Prashant (University of Georgia)
We present a new game-theoretic framework where Bayesian players engage in a Markov game and each has private but imperfect information regarding other players' types. Instead of utilizing Harsanyi's abstract types and a common prior distribution, we construct player types whose structure is explicit and induces a finite level belief hierarchy. We characterize equilibria in this game and formalize the computation of finding such equilibria as a constraint satisfaction problem. The effectiveness of the new framework is demonstrated on two ad hoc team work domains.
Individual Planning in Infinite-Horizon Multiagent Settings: Inference, Structure and Scalability
Qu, Xia, Doshi, Prashant
This paper provides the first formalization of self-interested planning in multiagent settings using expectation-maximization (EM). Our formalization in the context of infinite-horizon and finitely-nested interactive POMDPs (I-POMDP) is distinct from EM formulations for POMDPs and cooperative multiagent planning frameworks. We exploit the graphical model structure specific to I-POMDPs, and present a new approach based on block-coordinate descent for further speed up. Forward filtering-backward sampling -- a combination of exact filtering with sampling -- is explored to exploit problem structure.
Individual Planning in Agent Populations: Exploiting Anonymity and Frame-Action Hypergraphs
Sonu, Ekhlas, Chen, Yingke, Doshi, Prashant
Interactive partially observable Markov decision processes (I-POMDP) provide a formal framework for planning for a self-interested agent in multiagent settings. An agent operating in a multiagent environment must deliberate about the actions that other agents may take and the effect these actions have on the environment and the rewards it receives. Traditional I-POMDPs model this dependence on the actions of other agents using joint action and model spaces. Therefore, the solution complexity grows exponentially with the number of agents thereby complicating scalability. In this paper, we model and extend anonymity and context-specific independence -- problem structures often present in agent populations -- for computational gain. We empirically demonstrate the efficiency from exploiting these problem structures by solving a new multiagent problem involving more than 1,000 agents.
Monte Carlo Sampling Methods for Approximating Interactive POMDPs
Doshi, Prashant, Gmytrasiewicz, Piotr J.
Partially observable Markov decision processes (POMDPs) provide a principled framework for sequential planning in uncertain single agent settings. An extension of POMDPs to multiagent settings, called interactive POMDPs (I-POMDPs), replaces POMDP belief spaces with interactive hierarchical belief systems which represent an agent's belief about the physical world, about beliefs of other agents, and about their beliefs about others' beliefs. This modification makes the difficulties of obtaining solutions due to complexity of the belief and policy spaces even more acute. We describe a general method for obtaining approximate solutions of I-POMDPs based on particle filtering (PF). We introduce the interactive PF, which descends the levels of the interactive belief hierarchies and samples and propagates beliefs at each level. The interactive PF is able to mitigate the belief space complexity, but it does not address the policy space complexity. To mitigate the policy space complexity -- sometimes also called the curse of history -- we utilize a complementary method based on sampling likely observations while building the look ahead reachability tree. While this approach does not completely address the curse of history, it beats back the curse's impact substantially. We provide experimental results and chart future work.
Improved Convergence of Iterative Ontology Alignment using Block-Coordinate Descent
Thayasivam, Uthayasanker (University of Georgia) | Doshi, Prashant (University of Georgia)
A wealth of ontologies, many of which overlap in their scope, has made aligning ontologies an important problem for the semantic Web. Consequently, several algorithms now exist for automatically aligning ontologies, with mixed success in their performances. Crucial challenges for these algorithms involve scaling to large ontologies, and as applications of ontology alignment evolve, performing the alignment in a reasonable amount of time without compromising on the quality of the alignment. A class of alignment algorithms is iterative and often consumes more time than others while delivering solutions of high quality. We present a novel and general approach for speeding up the multivariable optimization process utilized by these algorithms. Specifically, we use the technique of block-coordinate descent in order to possibly improve the speed of convergence of the iterative alignment techniques. We integrate this approach into three well-known alignment systems and show that the enhanced systems generate similar or improved alignments in significantly less time on a comprehensive testbed of ontology pairs. This represents an important step toward making alignment techniques computationally more feasible.
Reports of the AAAI 2011 Conference Workshops
Agmon, Noa (University of Texas at Austin) | Agrawal, Vikas (Infosys Labs) | Aha, David W. (Naval Research Laboratory) | Aloimonos, Yiannis (University of Maryland, College Park) | Buckley, Donagh (EMC) | Doshi, Prashant (University of Georgia) | Geib, Christopher (University of Edinburgh) | Grasso, Floriana (University of Liverpool) | Green, Nancy (University of North Carolina Greensboro) | Johnston, Benjamin (University of Technology, Sydney) | Kaliski, Burt (VeriSign, Inc.) | Kiekintveld, Christopher (University of Texas at El Paso) | Law, Edith (Carnegie Mellon University) | Lieberman, Henry (Massachusetts Institute of Technology) | Mengshoel, Ole J. (Carnegie Mellon University) | Metzler, Ted (Oklahoma City University) | Modayil, Joseph (University of Alberta) | Oard, Douglas W. (University of Maryland, College Park) | Onder, Nilufer (Michigan Technological University) | O'Sullivan, Barry (University College Cork) | Pastra, Katerina (Cognitive Systems Research Insitute) | Precup, Doina (McGill University) | Ramachandran, Sowmya (Stottler Henke Associates, Inc.) | Reed, Chris (University of Dundee) | Sariel-Talay, Sanem (Istanbul Technical University) | Selker, Ted (Carnegie Mellon University) | Shastri, Lokendra (Infosys Technologies Ltd.) | Smith, Stephen F. (Carnegie Mellon University) | Singh, Satinder (University of Michigan at Ann Arbor) | Srivastava, Siddharth (University of Wisconsin, Madison) | Sukthankar, Gita (University of Central Florida) | Uthus, David C. (Naval Research Laboratory) | Williams, Mary-Anne (University of Technology, Sydney)
The AAAI-11 workshop program was held Sunday and Monday, August 7–18, 2011, at the Hyatt Regency San Francisco in San Francisco, California USA. The AAAI-11 workshop program included 15 workshops covering a wide range of topics in artificial intelligence. The titles of the workshops were Activity Context Representation: Techniques and Languages; Analyzing Microtext; Applied Adversarial Reasoning and Risk Modeling; Artificial Intelligence and Smarter Living: The Conquest of Complexity; AI for Data Center Management and Cloud Computing; Automated Action Planning for Autonomous Mobile Robots; Computational Models of Natural Argument; Generalized Planning; Human Computation; Human-Robot Interaction in Elder Care; Interactive Decision Theory and Game Theory; Language-Action Tools for Cognitive Artificial Agents: Integrating Vision, Action and Language; Lifelong Learning; Plan, Activity, and Intent Recognition; and Scalable Integration of Analytics and Visualization. This article presents short summaries of those events.