Goto

Collaborating Authors

 Country


Decentralised Metacognition in Context-Aware Autonomic Systems: Some Key Challenges

AAAI Conferences

A distributed non-hierarchical metacognitive architec- ture is one in which all meta-level reasoning compo- nents are subject to meta-level monitoring and manage- ment by other components. Such metacognitive distri- bution can support the robustness of distributed IT sys- tems in which humans and artificial agents are partic- ipants. However, robust metacognition also needs to be context-aware and use diversity in its reasoning and analysis methods. Both these requirements mean that an agent evaluates its reasoning within a “bigger picture” and that it can monitor this global picture from multi- ple perspectives. In particular, social context-awareness involves understanding the goals and concerns of users and organisations. In this paper, we first present a conceptual architecture for distributed metacognition with context-awareness and diversity. We then consider the challenges of apply- ing this architecture to autonomic management systems in scenarios where agents must collectively diagnose and respond to errors and intrusions. Such autonomic systems need rich semantic knowledge and diverse data sources in order to provide the necessary context for their metacognitive evaluations and decisions.


Hierarchical Planning for Mobile Manipulation

AAAI Conferences

Humans somehow manage to choose quite intelligently planner should fill in to produce a concrete plan that accomplishes the 20 trillion primitive motor commands that constitute a the goal as quickly as possible. It has long been thought that hierarchical structure in Planning at multiple levels of abstraction has long been a behavior is essential in managing this complexity. For instance, Shakey the exists at many levels, ranging from small (hundred-step?) robot used STRIPS for high-level task planning, then called motor programs for typing characters and saying phonemes out to separate low-level planning/control algorithms to execute up to large (billion-step?) actions such as writing an ICAPS each of the planned actions (Fikes and Nilsson 1971). This hard separation of levels, where a high-level plan is We believe that leveraging hierarchical structure will be chosen before considering low-level details, greatly simplifies equally important in achieving robust, efficient robotic behaviors. However, the resulting plans While your household robot probably won't get may be inefficient or even infeasible due to missed lowerlevel tenure anytime soon, even simple domestic tasks still have synergies and conflicts.


MCRNR: Fast Computing of Restricted Nash Responses by Means of Sampling

AAAI Conferences

This paper presents a sample-based algorithm for the computation of restricted Nash strategies in complex extensive form games. Recent work indicates that regret-minimization algorithms using selective sampling, such as Monte-Carlo Counterfactual Regret Minimization (MCCFR), converge faster to Nash-equilibrium (NE) strategies than their non-sampled counterparts which perform a full tree traversal. In this paper, we show that MCCFR is also able to establish NE strategies in the complex domain of Poker. Although such strategies are defensive (i.e. safe to play), they are oblivious to opponent mistakes. We can thus achieve better performance by using (an estimation of) opponent strategies. The Restricted Nash Response (RNR) algorithm was proposed to learn robust counter-strategies given such knowledge. It solves a modified game, wherein it is assumed that opponents play according to a fixed strategy with a certain probability, or to a regret-minimizing strategy otherwise. We improve the rate of convergence of the RNR algorithm using sampling. Our new algorithm, MCRNR, samples only relevant parts of the game tree. It is therefore able to converge faster to robust best-response strategies than RNR.We evaluate our algorithm on a variety of imperfect information games that are small enough to solve yet large enough to be strategically interesting, as well as a large game, Texas Hold’em Poker.


Parallel Best-First Search: The Role of Abstraction

AAAI Conferences

To harness modern multicore processors, it is imperative to develop parallel versions of fundamental algorithms. In this paper, we present a general approach to best-first heuristic search in a shared-memory setting. Each thread attempts to expand the most promising nodes. By using abstraction to partition the state space, we detect duplicate states while avoiding lock contention. We allow speculative expansions when necessary to keep threads busy. We identify and fix potential livelock conditions. In an empirical comparison on STRIPS planning, grid pathfinding, and sliding tile puzzle problems using an 8-core machine, we show that A* implemented in our framework yields faster search performance than previous parallel search proposals. We also demonstrate that our approach extends easily to other best-first searches, such as weighted A* and anytime heuristic search.


Using a Trust Model in Decision Making for Supply Chain Management

AAAI Conferences

One of the critical factors for a successful cooperative relationship in a supply chain partnership is trust. Many real-world applications, such as Supply Chain Management (SCM), can be modeled using multi-agent systems. One shortcoming of current SCM models is that their trust models are ad hoc and do not have a strong theoretical basis. As a result, they are unable to model subtleties in agent behavior that can be used to build a more accurate trust model. We propose a trust model for SCM that is grounded in probabilistic game theory. In this model, trust can be gained through direct interactions and/or by asking for information from other trustworthy agents. We will use this model to simulate and study supply chain market behavior.


A Cognitive Hierarchy Model Applied to the Lemonade Game

AAAI Conferences

One of the challenges of multiagent decision making is that the behavior needed to maximize utility can depend on what other agents choose to do: sometimes there is no "right" answer in the absence of knowledge of how opponents will act. The Nash equilibrium is a sensible choice of behavior because it represents a mutual best response. But, even when there is a unique equilibrium, other players are under no obligation to take part in it. This observation has been forcefully illustrated in the behavioral economics community where repeated experiments have shown individuals playing Nash equilibria and performing badly as a result. In this paper, we show how to apply a tool from behavioral economics called the Cognitive Hierarchy (CH) to the design of agents in general sum games. We attack the recently introduced ``Lemonade Game'' and show how the results of an open competition are well explained by CH. We believe this game, and perhaps many other similar games, boils down to predicting how deeply other agents in the game will be reasoning. An agent that does not reason enough risks being exploited by its opponents, while an agent that reasons too much may not be able to interact productively with its opponents. We demonstrate these ideas by presenting empirical results using agents from the competition and idealizations arising from a CH analysis.


Hierarchical Planning in the Now

AAAI Conferences

In this paper we outline an approach to the integration of task planning and motion planning that has the following key properties: It is aggressively hierarchical. It makes choices and commits to them in a top-down fashion in an attempt to limit the length of plans that need to be constructed, and thereby exponentially decrease the amount of search required. Importantly, our approach also limits the need to project the effect of actions into the far future. It operates on detailed, continuous geometric representations and partial symbolic descriptions. It does not require a complete symbolic representation of the input geometry or of the geometric effect of the task-level operations.


Efficient Lifting for Online Probabilistic Inference

AAAI Conferences

Lifting can greatly reduce the cost of inference on first-order probabilistic graphical models, but constructing the lifted network can itself be quite costly. In online applications (e.g., video segmentation) repeatedly constructing the lifted network for each new inference can be extremely wasteful, because the evidence typically changes little from one inference to the next. The same is true in many other problems that require repeated inference, like utility maximization, MAP inference, interactive inference, parameter and structure learning, etc. In this paper, we propose an efficient algorithm for updating the structure of an existing lifted network with incremental changes to the evidence. This allows us to construct the lifted network once for the initial inference problem, and amortize the cost over the subsequent problems. Experiments on video segmentation and viral marketing problems show that the algorithm greatly reduces the cost of inference without affecting the quality of the solutions.


Approaches for Automatically Enriching Wikipedia

AAAI Conferences

We have been exploring the use of Web-derived knowledge bases through the development of Wikitology — a hybrid knowledge base of structured and unstructured information extracted from Wikipedia augmented by RDF data from DBpedia and other Linked Open Data resources. In this paper, we describe approaches that aid in enriching Wikipedia and thus the resources that derive from Wikipedia such as the Wikitology knowledge base, DBpedia, Freebase and Powerset.


IsisWorld: An Open Source Commonsense Simulator for AI Researchers

AAAI Conferences

A metareasoning problem involves three parts: 1) a set of concrete problem domains; 2) reasoners to reason about the problems; and, 3) metareasoners to reason about the reasoners. We believe that the metareasoning community would benefit from agreeing on the first two problems. To support this kind of collaboration, we offer an open source 3D simulator containing everyday, commonsense problems that take place in kitchens. This paper presents several arguments for using a simulator to solve commonsense problems. The paper concludes by describing future work in simulator-based unified generative benchmarks for AI.