gnome
Generating Novelty in Open-World Multi-Agent Strategic Board Games
Kejriwal, Mayank, Thomas, Shilpa
We describe GNOME (Generating Novelty in Open-world Multi-agent Environments), an experimental platform that is designed to test the effectiveness of multi-agent AI systems when faced with \emph{novelty}. GNOME separates the development of AI gameplaying agents with the simulator, allowing \emph{unanticipated} novelty (in essence, novelty that is not subject to model-selection bias). Using a Web GUI, GNOME was recently demonstrated at NeurIPS 2020 using the game of Monopoly to foster an open discussion on AI robustness and the nature of novelty in real-world environments. In this article, we further detail the key elements of the demonstration, and also provide an overview of the experimental design that is being currently used in the DARPA Science of Artificial Intelligence and Learning for Open-World Novelty (SAIL-ON) program to evaluate external teams developing novelty-adaptive gameplaying agents.
- North America > United States > California > Monterey County > Marina (0.04)
- North America > United States > Connecticut (0.04)
Fast-Slow-Thinking: Complex Task Solving with Large Language Models
Sun, Yiliu, Zhang, Yanfang, Zhao, Zicheng, Wan, Sheng, Tao, Dacheng, Gong, Chen
Nowadays, Large Language Models (LLMs) have been gradually employed to solve complex tasks. To face the challenge, task decomposition has become an effective way, which proposes to divide a complex task into multiple simpler subtasks and then solve them separately so that the difficulty of the original task can be reduced. However, the performance of existing task decomposition methods can be suboptimal when the task contains overly complex logic and constraints. In this situation, the solution generated by LLMs may deviate from the original purpose of the task, or contain redundant or even erroneous content. Therefore, inspired by the fact that humans possess two thinking systems including fast thinking and slow thinking, this paper introduces a new task decomposition method termed ``Fast-Slow-Thinking'' (FST), which stimulates LLMs to solve tasks through the cooperation of Fast Thinking (FT) and Slow Thinking (ST) steps. Here FT focuses more on the general and concise aspect of the task, and ST focuses more on the details of the task. In FT, LLMs are prompted to remove the constraints of the original task, therefore simplifying it to a general and concise one. In ST, we recall the constraints removed in FT, so that LLMs can improve the answer generated in FT to meet the requirements of the original task. Therefore, our FST method enables LLMs to consider a complex problem via a human-like cognition process from coarse to fine, the effectiveness of which has been well demonstrated by the experiments on three types of tasks.
- Asia > China > Jiangsu Province > Nanjing (0.04)
- Oceania > Australia > New South Wales > Sydney (0.04)
- Europe > Romania > Sud - Muntenia Development Region > Giurgiu County > Giurgiu (0.04)
- Asia > Middle East > Jordan (0.04)
GNOME: Generating Negotiations through Open-Domain Mapping of Exchanges
Deshpande, Darshan, Sinha, Shambhavi, Kumar, Anirudh Ravi, Pal, Debaditya, May, Jonathan
Language Models have previously shown strong negotiation capabilities in closed domains where the negotiation strategy prediction scope is constrained to a specific setup. In this paper, we first show that these models are not generalizable beyond their original training domain despite their wide-scale pretraining. Following this, we propose an automated framework called GNOME, which processes existing human-annotated, closed-domain datasets using Large Language Models and produces synthetic open-domain dialogues for negotiation. GNOME improves the generalizability of negotiation systems while reducing the expensive and subjective task of manual data curation. Through our experimental setup, we create a benchmark comparing encoder and decoder models trained on existing datasets against datasets created through GNOME. Our results show that models trained on our dataset not only perform better than previous state of the art models on domain specific strategy prediction, but also generalize better to previously unseen domains.
- North America > United States > California (0.14)
- North America > Canada > Ontario > Toronto (0.04)
- Europe > Middle East > Malta > Eastern Region > Northern Harbour District > St. Julian's (0.04)
- (2 more...)
Human-Agent Cooperation in Games under Incomplete Information through Natural Language Communication
Chen, Shenghui, Fried, Daniel, Topcu, Ufuk
Developing autonomous agents that can strategize and cooperate with humans under information asymmetry is challenging without effective communication in natural language. We introduce a shared-control game, where two players collectively control a token in alternating turns to achieve a common objective under incomplete information. We formulate a policy synthesis problem for an autonomous agent in this game with a human as the other player. To solve this problem, we propose a communication-based approach comprising a language module and a planning module. The language module translates natural language messages into and from a finite set of flags, a compact representation defined to capture player intents. The planning module leverages these flags to compute a policy using an asymmetric information-set Monte Carlo tree search with flag exchange algorithm we present. We evaluate the effectiveness of this approach in a testbed based on Gnomes at Night, a search-and-find maze board game. Results of human subject experiments show that communication narrows the information gap between players and enhances human-agent cooperation efficiency with fewer turns.
- North America > United States > Texas > Travis County > Austin (0.04)
- North America > United States > Pennsylvania > Allegheny County > Pittsburgh (0.04)
The Download: abandoning carbon offsets, and creating new materials
In the fall of 2018, the University of California tasked a team of researchers with identifying projects from which it could confidently purchase carbon offsets that would reliably cancel out greenhouse gas emissions across its campuses. They found next to nothing. The findings helped prompt the entire university system to radically rethink its sustainability plans. Now the researchers are sharing the lessons they learned over the course of the project, in the hopes of helping other universities and organizations consider what role, if any, offsets should play in sustainability strategies, MIT Technology Review can report. The project's leaders have three main takeaways for what others should do.
- Government (0.64)
- Energy (0.42)
Google DeepMind AI Breakthrough Could Help Battery and Chip Development
Researchers at Google DeepMind have used artificial intelligence to predict the structures of more than 2 million new materials, in a breakthrough that could have wide-reaching benefits in sectors such as renewable energy and computing. DeepMind published 381,000 of the 2.2 million crystal structures that it predicts to be most stable. The breakthrough increases the number of known stable materials by a factor of ten. Although the materials will still need to be synthesized and tested, steps which can take months or even years, the latest development is expected to accelerate the discovery of new materials, which will be required for applications such as energy storage, solar cells, and superconductor chips. "While materials play a very critical role in almost any technology, we as humanity know only about a few tens of thousands of stable materials," says Ekin Dogus Cubuk, a Staff Research Scientist at Google Brain, who worked on the DeepMind AI tool, known as Graph Networks for Materials Exploration (GNoME).
- Energy > Renewable (0.56)
- Energy > Energy Storage (0.36)
Crystal-hunting DeepMind AI could help discover new wonder materials
A crystal structure predicted by the GNoME AI. It contains barium (blue), niobium (white) and oxygen (green). An artificial intelligence created by Google DeepMind may help revolutionise materials science, providing new ways to make better batteries, solar panels, computer chips and many more vital technologies. "Anytime somebody wants to improve their technology, it inevitably includes improving the materials," says Ekin Dogus Cubuk at DeepMind. "We just wanted them to have more options."
- North America > United States > California > Alameda County > Berkeley (0.07)
- Europe > United Kingdom > England > Merseyside > Liverpool (0.05)
- Europe > United Kingdom > England > Hampshire > Southampton (0.05)
An AI Dreamed Up 380,000 New Materials. The Next Challenge Is Making Them
The robotic line cooks were deep in their recipe, toiling away in a room tightly packed with equipment. In one corner, an articulated arm selected and mixed ingredients, while another slid back and forth on a fixed track, working the ovens. A third was on plating duty, carefully shaking the contents of a crucible onto a dish. Gerbrand Ceder, a materials scientist at Lawrence Berkeley Lab and UC Berkeley, nodded approvingly as a robotic arm delicately pinched and capped an empty plastic vial--an especially tricky task, and one of his favorites to observe. "These guys can work all night," Ceder said, giving two of his grad students a wry look.
Google DeepMind's new AI tool helped create more than 700 new materials
GNoME can be described as AlphaFold for materials discovery, according to Ju Li, a materials science and engineering professor at the Massachusetts Institute of Technology. AlphaFold, a DeepMind AI system announced in 2020, predicts the structures of proteins with high accuracy and has since advanced biological research and drug discovery. Thanks to GNoME, the number of known stable materials has grown almost tenfold, to 421,000. "While materials play a very critical role in almost any technology, we as humanity know only a few tens of thousands of stable materials," said Dogus Cubuk, materials discovery lead at Google DeepMind, at a press briefing. To discover new materials, scientists combine elements across the periodic table.
SCREWS: A Modular Framework for Reasoning with Revisions
Shridhar, Kumar, Jhamtani, Harsh, Fang, Hao, Van Durme, Benjamin, Eisner, Jason, Xia, Patrick
Large language models (LLMs) can improve their accuracy on various tasks through iteratively refining and revising their output based on feedback. We observe that these revisions can introduce errors, in which case it is better to roll back to a previous result. Further, revisions are typically homogeneous: they use the same reasoning method that produced the initial answer, which may not correct errors. To enable exploration in this space, we present SCREWS, a modular framework for reasoning with revisions. It is comprised of three main modules: Sampling, Conditional Resampling, and Selection, each consisting of sub-modules that can be hand-selected per task. We show that SCREWS not only unifies several previous approaches under a common framework, but also reveals several novel strategies for identifying improved reasoning chains. We evaluate our framework with state-of-the-art LLMs (ChatGPT and GPT-4) on a diverse set of reasoning tasks and uncover useful new reasoning strategies for each: arithmetic word problems, multi-hop question answering, and code debugging. Heterogeneous revision strategies prove to be important, as does selection between original and revised candidates.
- North America > Canada > Ontario > Toronto (0.04)
- Asia > Middle East > UAE > Abu Dhabi Emirate > Abu Dhabi (0.04)
- Europe > Switzerland > Zürich > Zürich (0.04)
- Europe > Italy > Tuscany > Florence (0.04)
- Research Report (0.64)
- Workflow (0.46)