Goto

Collaborating Authors

 diver


Learning to Dive in Branch and Bound

Neural Information Processing Systems

They iteratively modify and resolve linear programs to conduct a depth-first search from any node in the search tree. Existing divers rely on generic decision rules that fail to exploit structural commonality between similar problem instances that often arise in practice.


Inside the world's longest underwater cave: Subterranean water 'web' in Mexico extends at least 325 MILES

Daily Mail - Science & tech

Leaked recording reveals Campbell's exec's sickening remarks about iconic soup's ingredients How Lauren Sanchez would REALLY look if she'd never had rumored plastic surgery Trump's losing control... MAGA's imploding... and White House insiders tell me why they're REALLY worried: ANDREW NEIL Billionaire family posts VERY unusual obituary after heir, 40, met violent end at $2.8m hunting lodge following marriage scandal These women have lost as much as nine stone WITHOUT jabs: Now they reveal secret to their stunning success, the extraordinary event that brought them together and how it's changed their lives... Judge throws out Comey and James cases as Trump's beauty queen prosecutor is humiliated Her moving videos about the handsome boyfriend who ghosted her went viral and catapulted her to overnight fame. Kate Gosselin's ex Jon is seen at his splashy wedding for the first time as son Collin weighs in on his siblings not attending Fugitive'Slender Man' stabber Morgan Geyser snapped'just Google me' when asked for ID by cops who found her with MUCH older lover It all seems to be falling apart now! Pete Hegseth drops hammer on Democrat senator in'sedition' storm as court martial looms after Trump's execution threat Sabrina Carpenter looks unrecognisable in throwback snap from seven years ago as fans call her rebranding'wild' Neuralink's'Patient 4' feared missing months after getting revolutionary brain chip... now his wife tells the REAL heartbreaking story NFL's first transgender cheerleader makes explosive allegation against Carolina Panthers Slash your cholesterol by a third in just a month... hundreds of thousands are on a new diet that's transforming lives. Inside the world's longest underwater cave: Subterranean water'web' in Mexico extends at least 325 MILES Beneath the idyllic resort towns of Mexico's Yucatan Peninsula, daring explorers have uncovered a hidden world of grand chambers and twisting tunnels. The Ox Bel Ha, Mayan for'Three Paths of Water', is a sprawling water'web' that makes up the world's longest underwater cave system.


Adaptive GR(1) Specification Repair for Liveness-Preserving Shielding in Reinforcement Learning

Georgescu, Tiberiu-Andrei, Goodall, Alexander W., Alrajeh, Dalal, Belardinelli, Francesco, Uchitel, Sebastian

arXiv.org Artificial Intelligence

Shielding is widely used to enforce safety in reinforcement learning (RL), ensuring that an agent's actions remain compliant with formal specifications. Classical shielding approaches, however, are often static, in the sense that they assume fixed logical specifications and hand-crafted abstractions. While these static shields provide safety under nominal assumptions, they fail to adapt when environment assumptions are violated. In this paper, we develop the first adaptive shielding framework - to the best of our knowledge - based on Generalized Reactivity of rank 1 (GR(1)) specifications, a tractable and expressive fragment of Linear Temporal Logic (LTL) that captures both safety and liveness properties. Our method detects environment assumption violations at runtime and employs Inductive Logic Programming (ILP) to automatically repair GR(1) specifications online, in a systematic and interpretable way. This ensures that the shield evolves gracefully, ensuring liveness is achievable and weakening goals only when necessary. We consider two case studies: Minepump and Atari Seaquest; showing that (i) static symbolic controllers are often severely suboptimal when optimizing for auxiliary rewards, and (ii) RL agents equipped with our adaptive shield maintain near-optimal reward and perfect logical compliance compared with static shields.


AquaVLM: Improving Underwater Situation Awareness with Mobile Vision Language Models

Tian, Beitong, Zhao, Lingzhi, Chen, Bo, Zheng, Haozhen, Yang, Jingcheng, Wu, Mingyuan, Vasisht, Deepak, Nahrstedt, Klara

arXiv.org Artificial Intelligence

Underwater activities like scuba diving enable millions annually to explore marine environments for recreation and scientific research. Maintaining situational awareness and effective communication are essential for diver safety. Traditional underwater communication systems are often bulky and expensive, limiting their accessibility to divers of all levels. While recent systems leverage lightweight smartphones and support text messaging, the messages are predefined and thus restrict context-specific communication. In this paper, we present AquaVLM, a tap-and-send underwater communication system that automatically generates context-aware messages and transmits them using ubiquitous smartphones. Our system features a mobile vision-language model (VLM) fine-tuned on an auto-generated underwater conversation dataset and employs a hierarchical message generation pipeline. We co-design the VLM and transmission, incorporating error-resilient fine-tuning to improve the system's robustness to transmission errors. We develop a VR simulator to enable users to experience AquaVLM in a realistic underwater environment and create a fully functional prototype on the iOS platform for real-world experiments. Both subjective and objective evaluations validate the effectiveness of AquaVLM and highlight its potential for personal underwater communication as well as broader mobile VLM applications.


Robotic Classification of Divers' Swimming States using Visual Pose Keypoints as IMUs

Kutzke, Demetrious T., Wu, Ying-Kun, Terveen, Elizabeth, Sattar, Junaed

arXiv.org Artificial Intelligence

Traditional human activity recognition uses either direct image analysis or data from wearable inertial measurement units (IMUs), but can be ineffective in challenging underwater environments. We introduce a novel hybrid approach that bridges this gap to monitor scuba diver safety. Our method leverages computer vision to generate high-fidelity motion data, effectively creating a ``pseudo-IMU'' from a stream of 3D human joint keypoints. This technique circumvents the critical problem of wireless signal attenuation in water, which plagues conventional diver-worn sensors communicating with an Autonomous Underwater Vehicle (AUV). We apply this system to the vital task of identifying anomalous scuba diver behavior that signals the onset of a medical emergency such as cardiac arrest -- a leading cause of scuba diving fatalities. By integrating our classifier onboard an AUV and conducting experiments with simulated distress scenarios, we demonstrate the utility and effectiveness of our method for advancing robotic monitoring and diver safety.


Learning to Dive in Branch and Bound

Neural Information Processing Systems

They iteratively modify and resolve linear programs to conduct a depth-first search from any node in the search tree. Existing divers rely on generic decision rules that fail to exploit structural commonality between similar problem instances that often arise in practice.


One-Shot Gesture Recognition for Underwater Diver-To-Robot Communication

Joshi, Rishikesh, Sattar, Junaed

arXiv.org Artificial Intelligence

Reliable human-robot communication is essential for underwater human-robot interaction (U-HRI), yet traditional methods such as acoustic signaling and predefined gesture-based models suffer from limitations in adaptability and robustness. In this work, we propose One-Shot Gesture Recognition (OSG), a novel method that enables real-time, pose-based, temporal gesture recognition underwater from a single demonstration, eliminating the need for extensive dataset collection or model retraining. OSG leverages shape-based classification techniques, including Hu moments, Zernike moments, and Fourier descriptors, to robustly recognize gestures in visually-challenging underwater environments. Our system achieves high accuracy on real-world underwater data and operates efficiently on embedded hardware commonly found on autonomous underwater vehicles (AUVs), demonstrating its feasibility for deployment on-board robots. Compared to deep learning approaches, OSG is lightweight, computationally efficient, and highly adaptable, making it ideal for diver-to-robot communication. We evaluate OSG's performance on an augmented gesture dataset and real-world underwater video data, comparing its accuracy against deep learning methods. Our results show OSG's potential to enhance U-HRI by enabling the immediate deployment of user-defined gestures without the constraints of predefined gesture languages.


BlendRL: A Framework for Merging Symbolic and Neural Policy Learning

Shindo, Hikaru, Delfosse, Quentin, Dhami, Devendra Singh, Kersting, Kristian

arXiv.org Artificial Intelligence

Humans can leverage both symbolic reasoning and intuitive reactions. In contrast, reinforcement learning policies are typically encoded in either opaque systems like neural networks or symbolic systems that rely on predefined symbols and rules. This disjointed approach severely limits the agents' capabilities, as they often lack either the flexible low-level reaction characteristic of neural agents or the interpretable reasoning of symbolic agents. To overcome this challenge, we introduce BlendRL, a neuro-symbolic RL framework that harmoniously integrates both paradigms within RL agents that use mixtures of both logic and neural policies. We empirically demonstrate that BlendRL agents outperform both neural and symbolic baselines in standard Atari environments, and showcase their robustness to environmental changes. Additionally, we analyze the interaction between neural and symbolic policies, illustrating how their hybrid use helps agents overcome each other's limitations.


Comparative study of regression vs pairwise models for surrogate-based heuristic optimisation

Naharro, Pablo S., Toharia, Pablo, LaTorre, Antonio, Peña, José-María

arXiv.org Artificial Intelligence

Heuristic optimisation algorithms explore the search space by sampling solutions, evaluating their fitness, and biasing the search in the direction of promising solutions. However, in many cases, this fitness function involves executing expensive computational calculations, drastically reducing the reasonable number of evaluations. In this context, surrogate models have emerged as an excellent alternative to alleviate these computational problems. This paper addresses the formulation of surrogate problems as both regression models that approximate fitness (surface surrogate models) and a novel way to connect classification models (pairwise surrogate models). The pairwise approach can be directly exploited by some algorithms, such as Differential Evolution, in which the fitness value is not actually needed to drive the search, and it is sufficient to know whether a solution is better than another one or not. Based on these modelling approaches, we have conducted a multidimensional analysis of surrogate models under different configurations: different machine learning algorithms (regularised regression, neural networks, decision trees, boosting methods, and random forests), different surrogate strategies (encouraging diversity or relaxing prediction thresholds), and compare them for both surface and pairwise surrogate models. The experimental part of the article includes the benchmark problems already proposed for the SOCO2011 competition in continuous optimisation and a simulation problem included in the recent GECCO2021 Industrial Challenge. This paper shows that the performance of the overall search, when using online machine learning-based surrogate models, depends not only on the accuracy of the predictive model but also on both the kind of bias towards positive or negative cases and how the optimisation uses those predictions to decide whether to execute the actual fitness function.


OCALM: Object-Centric Assessment with Language Models

Kaufmann, Timo, Blüml, Jannis, Wüst, Antonia, Delfosse, Quentin, Kersting, Kristian, Hüllermeier, Eyke

arXiv.org Artificial Intelligence

Properly defining a reward signal to efficiently train a reinforcement learning (RL) agent is a challenging task. Designing balanced objective functions from which a desired behavior can emerge requires expert knowledge, especially for complex environments. Learning rewards from human feedback or using large language models (LLMs) to directly provide rewards are promising alternatives, allowing non-experts to specify goals for the agent. However, black-box reward models make it difficult to debug the reward. In this work, we propose Object-Centric Assessment with Language Models (OCALM) to derive inherently interpretable reward functions for RL agents from natural language task descriptions. OCALM uses the extensive world-knowledge of LLMs while leveraging the object-centric nature common to many environments to derive reward functions focused on relational concepts, providing RL agents with the ability to derive policies from task descriptions.