raven
Chris Pratt on new film Mercy: I asked to be locked into an executioner's chair
Chris Pratt on new film Mercy: I asked to be locked into an executioner's chair Being locked barefoot in an executioner's chair sounds uncomfortable, but that is what Chris Pratt requested for his latest film, Mercy. More familiar as a wisecracking action hero in blockbusters like Guardians of the Galaxy and Jurassic World, this role is quite a departure for him. He plays homicide detective Chris Raven, who's fighting for his life after being accused of murdering his wife. Raven is an alcoholic who wakes in the chair after a drinking binge, with just 90 minutes to convince an AI judge he's innocent, or he'll be executed immediately. The film is set in real time, so we see Raven defend his case - while enduring a crashing hangover.
- North America > United States (0.15)
- North America > Central America (0.15)
- Oceania > Australia (0.05)
- (15 more...)
- Media > Film (1.00)
- Leisure & Entertainment (1.00)
- Law Enforcement & Public Safety > Crime Prevention & Enforcement (1.00)
- (2 more...)
Video Models Start to Solve Chess, Maze, Sudoku, Mental Rotation, and Raven' Matrices
We show that video generation models could reason now. Testing on tasks such as chess, maze, Sudoku, mental rotation, and Raven's Matrices, leading models such as Sora-2 achieve sixty percent success rates. We establish a robust experimental paradigm centered on the "Task Pair" design. We build a code framework, with 39 models available already, that supports this paradigm and allows for easy scaling - users can add models and tasks efficiently. We show our automated evaluation strongly correlates with human judgment, and therefore this paradigm is highly scalable. We see an opportunity, given the availability of our paradigm, to do reinforcement learning for improving reasoning in video models. You could checkout all of our raw $\href{https://grow-ai-like-a-child.com/video-reason/}{results}$ and our $\href{https://github.com/hokindeng/VMEvalKit}{VMEvalKit}$ codebase.
- North America > United States > Pennsylvania > Allegheny County > Pittsburgh (0.04)
- North America > United States > Massachusetts > Middlesex County > Cambridge (0.04)
- Leisure & Entertainment > Games > Chess (0.88)
- Leisure & Entertainment > Games > Sudoku (0.63)
- Information Technology > Artificial Intelligence > Representation & Reasoning (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.70)
- Information Technology > Artificial Intelligence > Cognitive Science > Problem Solving (0.69)
- North America > United States > Missouri > Jackson County > Kansas City (0.48)
- North America > United States > Illinois > Cook County > Chicago (0.26)
- North America > United States > Maryland > Baltimore (0.05)
- (2 more...)
- Information Technology > Communications > Social Media (0.75)
- Information Technology > Artificial Intelligence > Games (0.73)
Author Philip Pullman calls on government to act on AI using books for training
Author Philip Pullman calls on government to act over'wicked' AI scraping Writers whose work has been scraped don't get compensation or recognition, something authors including Kate Mosse and Richard Osman have criticised, saying it could destroy growth in creative fields and amount to theft. Sir Philip, author of the hugely popular novels about Lyra Silvertongue, the heroine of His Dark Materials and The Book of Dust trilogies, thinks writers should be compensated. They can do what they like with my work if they pay me for it, he told the BBC's culture editor Katie Razzall. The Department for Culture, Media and Sport has been contacted for a response to Sir Philip's comments. Sir Philip said: As far as I know everybody's work has been stolen, scraped like a trawler... at the bottom of the sea. You name it, it's all killed.
- Oceania > Australia (0.18)
- North America > United States (0.17)
- South America (0.15)
- (13 more...)
RAVEN: Resilient Aerial Navigation via Open-Set Semantic Memory and Behavior Adaptation
Kim, Seungchan, Alama, Omar, Kurdydyk, Dmytro, Keller, John, Keetha, Nikhil, Wang, Wenshan, Bisk, Yonatan, Scherer, Sebastian
Aerial outdoor semantic navigation requires robots to explore large, unstructured environments to locate target objects. Recent advances in semantic navigation have demonstrated open-set object-goal navigation in indoor settings, but these methods remain limited by constrained spatial ranges and structured layouts, making them unsuitable for long-range outdoor search. While outdoor semantic navigation approaches exist, they either rely on reactive policies based on current observations, which tend to produce short-sighted behaviors, or precompute scene graphs offline for navigation, limiting adaptability to online deployment. We present RAVEN, a 3D memory-based, behavior tree framework for aerial semantic navigation in unstructured outdoor environments. It (1) uses a spatially consistent semantic voxel-ray map as persistent memory, enabling long-horizon planning and avoiding purely reactive behaviors, (2) combines short-range voxel search and long-range ray search to scale to large environments, (3) leverages a large vision-language model to suggest auxiliary cues, mitigating sparsity of outdoor targets. These components are coordinated by a behavior tree, which adaptively switches behaviors for robust operation. We evaluate RAVEN in 10 photorealistic outdoor simulation environments over 100 semantic tasks, encompassing single-object search, multi-class, multi-instance navigation and sequential task changes. Results show RAVEN outperforms baselines by 85.25% in simulation and demonstrate its real-world applicability through deployment on an aerial robot in outdoor field tests.
- Europe > Netherlands > South Holland > Delft (0.04)
- North America > United States > Pennsylvania > Allegheny County > Pittsburgh (0.04)
- Health & Medicine > Consumer Health (0.40)
- Transportation (0.31)
Flying robot leaps upwards and then takes to the air like a bird
A robot that can jump into flight like a bird could eliminate the need for runways for small fixed-winged drones. Birds use the powerful explosive force generated by their legs to leap into the air and start flying, but building a robot that can withstand the strong acceleration and forces involved in doing that has proved difficult. Now, Won Dong Shin at the Swiss Federal Technology Institute of Lausanne (EPFL) and his colleagues have built a flying propellered robot called RAVEN that can walk, hop and jump into the air to start flying, with legs that work like a bird's. "Fixed-wing vehicles, like airplanes, always require a runway or a launcher, which is not found everywhere. It really requires designated infrastructure to make an airplane take off," says Shin. "But if you see a bird, they just walk around, jump and take off. They don't need any external assistance."
Fast ground-to-air transition with avian-inspired multifunctional legs
Shin, Won Dong, Phan, Hoang-Vu, Daley, Monica A., Ijspeert, Auke J., Floreano, Dario
Most birds can navigate seamlessly between aerial and terrestrial environments. Whereas the forelimbs evolved into wings primarily for flight, the hindlimbs serve diverse functions such as walking, hopping, and leaping, and jumping take-off for transitions into flight. These capabilities have inspired engineers to aim for similar multi-modality in aerial robots, expanding their range of applications across diverse environments. However, challenges remain in reproducing multi-modal locomotion, across gaits with distinct kinematics and propulsive characteristics, such as walking and jumping, while preserving lightweight mass for flight. This tradeoff between mechanical complexity and versatility limits most existing aerial robots to only one additional locomotor mode. Here, we overcome the complexity-versatility tradeoff with RAVEN (Robotic Avian-inspired Vehicle for multiple ENvironments), which uses its bird-inspired multi-functional legs to jump rapidly into flight, walk on ground and hop over obstacles and gaps similar to the multi-modal locomotion of birds. We show that jumping for take-off contributes substantially to initial flight take-off speed and, remarkably, that it is more energy-efficient than solely propeller-based take-off. Our analysis suggests an important tradeoff in mass distribution between legs and body among birds adapted for different locomotor strategies, with greater investment in leg mass among terrestrial birds with multi-modal gait demands. Multi-functional robot legs expand opportunities to deploy traditional fixed-wing aircraft in complex terrains through autonomous take-offs and multi-modal gaits.
- North America > United States > California > Orange County > Irvine (0.14)
- Europe > Switzerland > Vaud > Lausanne (0.04)
- North America > Canada > British Columbia > Metro Vancouver Regional District > Vancouver (0.04)
- (4 more...)
- Energy (0.93)
- Water & Waste Management > Solid Waste Management (0.46)
Crowd IQ -- Aggregating Opinions to Boost Performance
Kosinski, Michal, Bachrach, Yoram, Graepel, Thore, Kasneci, Giergji, Van Gael, Jurgen
We show how the quality of decisions based on the aggregated opinions of the crowd can be conveniently studied using a sample of individual responses to a standard IQ questionnaire. We aggregated the responses to the IQ questionnaire using simple majority voting and a machine learning approach based on a probabilistic graphical model. The score for the aggregated questionnaire, Crowd IQ, serves as a quality measure of decisions based on aggregating opinions, which also allows quantifying individual and crowd performance on the same scale. We show that Crowd IQ grows quickly with the size of the crowd but saturates, and that for small homogeneous crowds the Crowd IQ significantly exceeds the IQ of even their most intelligent member. We investigate alternative ways of aggregating the responses and the impact of the aggregation method on the resulting Crowd IQ. We also discuss Contextual IQ, a method of quantifying the individual participant's contribution to the Crowd IQ based on the Shapley value from cooperative game theory.
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.04)
- Europe > Spain > Valencian Community > Valencia Province > Valencia (0.04)
- Education (0.46)
- Government (0.46)
- Leisure & Entertainment (0.34)
RAVEN: Multitask Retrieval Augmented Vision-Language Learning
Rao, Varun Nagaraj, Choudhary, Siddharth, Deshpande, Aditya, Satzoda, Ravi Kumar, Appalaraju, Srikar
The scaling of large language models to encode all the world's knowledge in model parameters is unsustainable and has exacerbated resource barriers. Retrieval-Augmented Generation (RAG) presents a potential solution, yet its application to vision-language models (VLMs) is under explored. Existing methods focus on models designed for single tasks. Furthermore, they're limited by the need for resource intensive pre training, additional parameter requirements, unaddressed modality prioritization and lack of clear benefit over non-retrieval baselines. This paper introduces RAVEN, a multitask retrieval augmented VLM framework that enhances base VLMs through efficient, task specific fine-tuning. By integrating retrieval augmented samples without the need for additional retrieval-specific parameters, we show that the model acquires retrieval properties that are effective across multiple tasks. Our results and extensive ablations across retrieved modalities for the image captioning and VQA tasks indicate significant performance improvements compared to non retrieved baselines +1 CIDEr on MSCOCO, +4 CIDEr on NoCaps and nearly a +3\% accuracy on specific VQA question types. This underscores the efficacy of applying RAG approaches to VLMs, marking a stride toward more efficient and accessible multimodal learning.
A Feature-based Generalizable Prediction Model for Both Perceptual and Abstract Reasoning
Do, Quan, Morin, Thomas M., Stern, Chantal E., Hasselmo, Michael E.
A hallmark of human intelligence is the ability to infer abstract rules from limited experience and apply these rules to unfamiliar situations. This capacity is widely studied in the visual domain using the Raven's Progressive Matrices. Recent advances in deep learning have led to multiple artificial neural network models matching or even surpassing human performance. However, while humans can identify and express the rule underlying these tasks with little to no exposure, contemporary neural networks often rely on massive pattern-based training and cannot express or extrapolate the rule inferred from the task. Furthermore, most Raven's Progressive Matrices or Raven-like tasks used for neural network training used symbolic representations, whereas humans can flexibly switch between symbolic and continuous perceptual representations. In this work, we present an algorithmic approach to rule detection and application using feature detection, affine transformation estimation and search. We applied our model to a simplified Raven's Progressive Matrices task, previously designed for behavioral testing and neuroimaging in humans. The model exhibited one-shot learning and achieved near human-level performance in the symbolic reasoning condition of the simplified task. Furthermore, the model can express the relationships discovered and generate multi-step predictions in accordance with the underlying rule. Finally, the model can reason using continuous patterns. We discuss our results and their relevance to studying abstract reasoning in humans, as well as their implications for improving intelligent machines.
- North America > United States > Massachusetts > Suffolk County > Boston (0.05)
- North America > United States > New York > New York County > New York City (0.04)
- North America > United States > Massachusetts > Middlesex County > Waltham (0.04)