trash
- Europe > Germany > Hesse > Darmstadt Region > Darmstadt (0.05)
- North America > Canada (0.04)
Choreographing Trash Cans: On Speculative Futures of Weak Robots in Public Spaces
Axelsson, Minja, Sikau, Lea Luka
Michio Okada first conceptualised "weak robots", which have limited capabilities themselves, and are framed as objects or "social others" which people are invited to assist and take care of. In Okada's work, such robots are used to invite pro-social behaviour from people, such as encouraging them to pick up trash to assist a trash can robot (Okada (2022)). We conceptualise human-robot interaction (HRI) as a stage where weak robots-- designed to be "cute" and vulnerable--play the role of incidental actors that subvert the person engaging with them. Caudwell and Lacey (2020) argue that cuteness as a design choice for robots can encourage users to trust and form relationships with those robots, which introduces ambivalent power dynamics through the production of intimacy . In fact, cuteness can also be seen as a deceptive or "dark" pattern, due to the utilisation of cuteness to prompt affective responses which can be used to collect emotional data, as well as some degree of reduction of user agency (Lacey and Caudwell (2019)). The ability and affordances of cute and weak robots to influence user behaviour merits the discussion of their ethicality, which we do in this paper through design fiction. Unlike traditional HRI research, often confined to laboratory settings, our focus is on spontaneous, real-world interactions that transform everyday environments into sites of performative potential. We argue that the theatricality of these encounters is central to understanding their impact: the presence of a weak and/or cute robot, such as the trash can robot, developed by Okada and the Interaction and Communication Design Lab of the T oyohashi University of T echnology, acts as a disruptive interloper that introduces an observer's effect and, thus, affects the human interlocutors. First, we examine the concept of weak robots through the lens of performativity theory as well as concepts of machine (dys)function.
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.14)
- North America > United States > Minnesota (0.04)
- North America > United States > Indiana (0.04)
Introspection in Learned Semantic Scene Graph Localisation
Bissessur, Manshika Charvi, Panagiotaki, Efimia, De Martini, Daniele
This work investigates how semantics influence localisation performance and robustness in a learned self-supervised, contrastive semantic localisation framework. After training a localisation network on both original and perturbed maps, we conduct a thorough post-hoc introspection analysis to probe whether the model filters environmental noise and prioritises distinctive landmarks over routine clutter. We validate various interpretability methods and present a comparative reliability analysis. Integrated gradients and Attention Weights consistently emerge as the most reliable probes of learned behaviour. A semantic class ablation further reveals an implicit weighting in which frequent objects are often down-weighted. Overall, the results indicate that the model learns noise-robust, semantically salient relations about place definition, thereby enabling explainable registration under challenging visual and structural variations.
- Europe > United Kingdom > England > Oxfordshire > Oxford (0.14)
- Europe > Switzerland (0.04)
- Asia > Singapore (0.04)
- (2 more...)
- Information Technology > Artificial Intelligence > Representation & Reasoning (1.00)
- Information Technology > Artificial Intelligence > Machine Learning (1.00)
- Information Technology > Artificial Intelligence > Vision (0.94)
- Information Technology > Artificial Intelligence > Natural Language > Text Processing (0.91)
Diverse Image Captioning with Context Object Split Latent Spaces
The word dimension for the embedding layer is 300. In Tab. 7 we further evaluate the diversity of COS-CVAE using self-CIDEr We provide additional qualitative results in Tabs. In Tab. 12 we show the divserse captions for novel objects generated by our model and the regions The evaluation server for nocaps accepts only one caption per image and does not support methods modeling one-to-many relationships for images and captions. In Figure 1 (left) we show the average accuracy and diversity scores again averaged across annotators; in Figure 1 (right) we show the accuracy and diversity scores from each annotator. We find that the captions generated by the COS-CV AE are scored to be more accurate compared to COS-CV AE (paired).
- Europe > Germany > Hesse > Darmstadt Region > Darmstadt (0.05)
- North America > Canada (0.04)
Mount Everest has a poo problem. Are drones the answer?
Breakthroughs, discoveries, and DIY tips sent every weekday. For some adventurers, scaling Mount Everest represents the ultimate test of grit and determination: a visual signifier of humanity's epic struggle to overcome the elements. For others, the peak can seem more like a really tall trash can. Every year, around 600 climbers make the trek from the mountain's base camp to the summit. During their time on Everest, each person produces an estimated 18 pounds of waste, most of which is left behind.
- Asia > Nepal (0.20)
- North America > United States > New York (0.05)
Adapting by Analogy: OOD Generalization of Visuomotor Policies via Functional Correspondence
Gupta, Pranay, Admoni, Henny, Bajcsy, Andrea
End-to-end visuomotor policies trained using behavior cloning have shown a remarkable ability to generate complex, multi-modal low-level robot behaviors. However, at deployment time, these policies still struggle to act reliably when faced with out-of-distribution (OOD) visuals induced by objects, backgrounds, or environment changes. Prior works in interactive imitation learning solicit corrective expert demonstrations under the OOD conditions -- but this can be costly and inefficient. We observe that task success under OOD conditions does not always warrant novel robot behaviors. In-distribution (ID) behaviors can directly be transferred to OOD conditions that share functional similarities with ID conditions. For example, behaviors trained to interact with in-distribution (ID) pens can apply to interacting with a visually-OOD pencil. The key challenge lies in disambiguating which ID observations functionally correspond to the OOD observation for the task at hand. We propose that an expert can provide this OOD-to-ID functional correspondence. Thus, instead of collecting new demonstrations and re-training at every OOD encounter, our method: (1) detects the need for feedback by first checking if current observations are OOD and then identifying whether the most similar training observations show divergent behaviors, (2) solicits functional correspondence feedback to disambiguate between those behaviors, and (3) intervenes on the OOD observations with the functionally corresponding ID observations to perform deployment-time generalization. We validate our method across diverse real-world robotic manipulation tasks with a Franka Panda robotic manipulator. Our results show that test-time functional correspondences can improve the generalization of a vision-based diffusion policy to OOD objects and environment conditions with low feedback.
- North America > United States > Pennsylvania > Allegheny County > Pittsburgh (0.04)
- North America > United States > Florida > Hillsborough County > University (0.04)
Things Are Getting More Expensive. There's an Easy Way to Save a Lot of Money.
Sign up for the Slatest to get the most insightful analysis, criticism, and advice out there, delivered to your inbox daily. Americans are mad as hell about high food prices. They hate paying more at the supermarket even more than they hate paying more at the pump. Food inflation was arguably their main reason for President Donald Trump's win, and Trump's failure to reverse it (while imposing tariffs that accelerate it) is arguably the main reason for his sinking approval ratings. Cost-conscious consumers have been clipping more coupons, dining out less, buying more generic brands, and generally changing their grocery shopping habits to save money.
- North America > United States > Ohio (0.07)
- North America > United States > California (0.05)
- Asia > China (0.05)
Making Sense of Robots in Public Spaces: A Study of Trash Barrel Robots
Bu, Fanjun, Fischer, Kerstin, Ju, Wendy
In this work, we analyze video data and interviews from a public deployment of two trash barrel robots in a large public space to better understand the sensemaking activities people perform when they encounter robots in public spaces. Based on an analysis of 274 human-robot interactions and interviews with N=65 individuals or groups, we discovered that people were responding not only to the robots or their behavior, but also to the general idea of deploying robots as trashcans, and the larger social implications of that idea. They wanted to understand details about the deployment because having that knowledge would change how they interact with the robot. Based on our data and analysis, we have provided implications for design that may be topics for future human-robot design researchers who are exploring robots for public space deployment. Furthermore, our work offers a practical example of analyzing field data to make sense of robots in public spaces.
- North America > United States > New York > New York County > New York City (0.14)
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.14)
- North America > United States > California > San Francisco County > San Francisco (0.04)
- (17 more...)
- Law (1.00)
- Health & Medicine (1.00)
- Government > Regional Government > North America Government > United States Government (0.46)
How to free up space on your Mac
Apple has made it easier than ever to merge duplicate photos. Are you tired of scrolling through your Mac's photo library only to find multiple copies of the same photo? Duplicate photos can clutter your storage and make it harder to find the memories you want to cherish. Fortunately, if you're using macOS Ventura or later, Apple has made it easier than ever to find and merge these duplicates right within the Photos app. We'll walk you through how to use the built-in Duplicates finder, as well as some alternative methods for those who need more advanced features.
- Information Technology > Security & Privacy (0.34)
- Media > News (0.31)
Scalable, Training-Free Visual Language Robotics: A Modular Multi-Model Framework for Consumer-Grade GPUs
Samson, Marie, Muraccioli, Bastien, Kanehiro, Fumio
The integration of language instructions with robotic control, particularly through Vision Language Action (VLA) models, has shown significant potential. However, these systems are often hindered by high computational costs, the need for extensive retraining, and limited scalability, making them less accessible for widespread use. In this paper, we introduce SVLR (Scalable Visual Language Robotics), an open-source, modular framework that operates without the need for retraining, providing a scalable solution for robotic control. SVLR leverages a combination of lightweight, open-source AI models including the Vision-Language Model (VLM) Mini-InternVL, zero-shot image segmentation model CLIPSeg, Large Language Model Phi-3, and sentence similarity model all-MiniLM to process visual and language inputs. These models work together to identify objects in an unknown environment, use them as parameters for task execution, and generate a sequence of actions in response to natural language instructions. A key strength of SVLR is its scalability. The framework allows for easy integration of new robotic tasks and robots by simply adding text descriptions and task definitions, without the need for retraining. This modularity ensures that SVLR can continuously adapt to the latest advancements in AI technologies and support a wide range of robots and tasks. SVLR operates effectively on an NVIDIA RTX 2070 (mobile) GPU, demonstrating promising performance in executing pick-and-place tasks. While these initial results are encouraging, further evaluation across a broader set of tasks and comparisons with existing VLA models are needed to assess SVLR's generalization capabilities and performance in more complex scenarios.