action program
- North America > United States > Ohio (0.04)
- North America > United States > Indiana > Monroe County > Bloomington (0.04)
- North America > Canada (0.04)
- North America > United States > Ohio (0.04)
- North America > United States > Indiana > Monroe County > Bloomington (0.04)
- North America > Canada > British Columbia > Metro Vancouver Regional District > Vancouver (0.04)
Emergent Braitenberg-style Behaviours for Navigating the ViZDoom `My Way Home' Labyrinth
Bayer, Caleidgh, Smith, Robert J., Heywood, Malcolm I.
The navigation of complex labyrinths with tens of rooms under visual partially observable state is typically addressed using recurrent deep reinforcement learning architectures. In this work, we show that navigation can be achieved through the emergent evolution of a simple Braitentberg-style heuristic that structures the interaction between agent and labyrinth, i.e. complex behaviour from simple heuristics. To do so, the approach of tangled program graphs is assumed in which programs cooperatively coevolve to develop a modular indexing scheme that only employs 0.8\% of the state space. We attribute this simplicity to several biases implicit in the representation, such as the use of pixel indexing as opposed to deploying a convolutional kernel or image processing operators.
LEAP: LLM-Generation of Egocentric Action Programs
Dessalene, Eadom, Maynord, Michael, Fermüller, Cornelia, Aloimonos, Yiannis
We introduce LEAP (illustrated in Figure 1), a novel method for generating video-grounded action programs through use of a Large Language Model (LLM). These action programs represent the motoric, perceptual, and structural aspects of action, and consist of sub-actions, pre-and post-conditions, and control flows. LEAP's action programs are centered on egocentric video and employ recent developments in LLMs both as a source for program knowledge and as an aggregator and assessor of multimodal video information. We apply LEAP over a majority (87%) of the training set of the EPIC Kitchens dataset, and release the resulting action programs as a publicly available dataset here. We employ LEAP as a secondary source of supervision, using its action programs in a loss term applied to action recognition and anticipation networks. We demonstrate sizable improvements in performance in both tasks due to training with the LEAP dataset. Our method achieves 1st place on the EPIC Kitchens Action Recognition leaderboard as of November 17 among the networks restricted to RGB-input (see Supplementary Materials).
Bongard-LOGO: A New Benchmark for Human-Level Concept Learning and Reasoning
Nie, Weili, Yu, Zhiding, Mao, Lei, Patel, Ankit B., Zhu, Yuke, Anandkumar, Animashree
Humans have an inherent ability to learn novel concepts from only a few samples and generalize these concepts to different situations. Even though today's machine learning models excel with a plethora of training data on standard recognition tasks, a considerable gap exists between machine-level pattern recognition and human-level concept learning. To narrow this gap, the Bongard Problems (BPs) were introduced as an inspirational challenge for visual cognition in intelligent systems. Despite new advances in representation learning and learning to learn, BPs remain a daunting challenge for modern AI. Inspired by the original one hundred BPs, we propose a new benchmark Bongard-LOGO for human-level concept learning and reasoning. We develop a program-guided generation technique to produce a large set of human-interpretable visual cognition problems in action-oriented LOGO language. Our benchmark captures three core properties of human cognition: 1) context-dependent perception, in which the same object may have disparate interpretations given different contexts; 2) analogy-making perception, in which some meaningful concepts are traded off for other meaningful concepts; and 3) perception with a few samples but infinite vocabulary. In experiments, we show that the state-of-the-art deep learning methods perform substantially worse than human subjects, implying that they fail to capture core human cognition properties. Finally, we discuss research directions towards a general architecture for visual reasoning to tackle this benchmark.
- North America > United States > Ohio (0.04)
- North America > Canada > British Columbia > Metro Vancouver Regional District > Vancouver (0.04)
Google turns product searches into cash by partnering with retailers
Google routinely fields product queries from millions of shoppers. Now, it wants to take a cut of their purchases, too. Under a new program, retailers can list their products on Google Search, as well as on the Google Express shopping service and Google Assistant. In exchange for Google listings and linking to retailer loyalty programs, the retailers pay Google a piece of each purchase, which is different from payments that retailers make to place ads on Google platforms. A new Google initiative called'Shopping Actions,' should help retailers fend off Amazon's growing dominance.
- Retail (1.00)
- Information Technology > Services (0.73)