Goto

Collaborating Authors

 san diego


ATLAS: Constraints-Aware Multi-Agent Collaboration for Real-World Travel Planning

Choi, Jihye, Yoon, Jinsung, Chen, Jiefeng, Jha, Somesh, Pfister, Tomas

arXiv.org Artificial Intelligence

While Large Language Models (LLMs) have shown remarkable advancements in reasoning and tool use, they often fail to generate optimal, grounded solutions under complex constraints. Real-world travel planning exemplifies these challenges, evaluating agents' abilities to handle constraints that are explicit, implicit, and even evolving based on interactions with dynamic environments and user needs. In this paper, we present ATLAS, a general multi-agent framework designed to effectively handle such complex nature of constraints awareness in real-world travel planning tasks. ATLAS introduces a principled approach to address the fundamental challenges of constraint-aware planning through dedicated mechanisms for dynamic constraint management, iterative plan critique, and adaptive interleaved search. ATLAS demonstrates state-of-the-art performance on the TravelPlanner benchmark, improving the final pass rate from 23.3% to 44.4% over its best alternative. More importantly, our work is the first to demonstrate quantitative effectiveness on real-world travel planning tasks with live information search and multi-turn feedback. In this realistic setting, ATLAS showcases its superior overall planning performance, achieving an 84% final pass rate which significantly outperforms baselines including ReAct (59%) and a monolithic agent (27%).


Taming Unbalanced Training Workloads in Deep Learning with Partial Collective Operations

Li, Shigang, Ben-Nun, Tal, Di Girolamo, Salvatore, Alistarh, Dan, Hoefler, Torsten

arXiv.org Artificial Intelligence

Load imbalance pervasively exists in distributed deep learning training systems, either caused by the inherent imbalance in learned tasks or by the system itself. Traditional synchronous Stochastic Gradient Descent (SGD) achieves good accuracy for a wide variety of tasks, but relies on global synchronization to accumulate the gradients at every training step. In this paper, we propose eager-SGD, which relaxes the global synchronization for decentralized accumulation. To implement eager-SGD, we propose to use two partial collectives: solo and majority. With solo allreduce, the faster processes contribute their gradients eagerly without waiting for the slower processes, whereas with majority allreduce, at least half of the participants must contribute gradients before continuing, all without using a central parameter server. We theoretically prove the convergence of the algorithms and describe the partial collectives in detail. Experimental results on load-imbalanced environments (CIFAR-10, ImageNet, and UCF101 datasets) show that eager-SGD achieves 1.27x speedup over the state-of-the-art synchronous SGD, without losing accuracy.


Masked home invader 'shot' after 'pistol-whipping' OnlyFans star, demanding cryptocurrency

FOX News

Fox News Flash top headlines are here. Check out what's clicking on Foxnews.com. A popular internet personality live-posted her own violent home invasion as a group of armed men stormed her home and demanded access to her cryptocurrency accounts. Video game streamer and adult content creator Kaitlyn Siragusa, who goes by the online name Amouranth, was asleep in her Houston home when three men shot through a patio window on Sunday evening, authorities told FOX 26. "I'm being too robbed at gunpoint," Siragusa posted on her X account.


Predicting Quality of Video Gaming Experience Using Global-Scale Telemetry Data and Federated Learning

Zhang, Zhongyang, Wen, Jinhe, Chen, Zixi, Arbab, Dara, Sahani, Sruti, Lewis, William, Giard, Kent, Arbab, Bijan, Jin, Haojian, Rahman, Tauhidur

arXiv.org Artificial Intelligence

Frames Per Second (FPS) significantly affects the gaming experience. Providing players with accurate FPS estimates prior to purchase benefits both players and game developers. However, we have a limited understanding of how to predict a game's FPS performance on a specific device. In this paper, we first conduct a comprehensive analysis of a wide range of factors that may affect game FPS on a global-scale dataset to identify the determinants of FPS. This includes player-side and game-side characteristics, as well as country-level socio-economic statistics. Furthermore, recognizing that accurate FPS predictions require extensive user data, which raises privacy concerns, we propose a federated learning-based model to ensure user privacy. Each player and game is assigned a unique learnable knowledge kernel that gradually extracts latent features for improved accuracy. We also introduce a novel training and prediction scheme that allows these kernels to be dynamically plug-and-play, effectively addressing cold start issues. To train this model with minimal bias, we collected a large telemetry dataset from 224 countries and regions, 100,000 users, and 835 games. Our model achieved a mean Wasserstein distance of 0.469 between predicted and ground truth FPS distributions, outperforming all baseline methods.


Why Surgeons Are Wearing The Apple Vision Pro In Operating Rooms

TIME - Tech

Twenty-four years ago, the surgeon Santiago Horgan performed the first robotically assisted gastric-bypass surgery in the world, a major medical breakthrough. Now Horgan is working with a new tool that he argues could be even more transformative in operating rooms: the Apple Vision Pro. Over the last month, Horgan and other surgeons at the University of California, San Diego have performed more than 20 minimally invasive operations while wearing Apple's mixed-reality headsets. Apple released the headsets to the public in February, and they've largely been a commercial flop. But practitioners in some industries, including architecture and medicine, have been testing how they might serve particular needs.


Sphere Neural-Networks for Rational Reasoning

Dong, Tiansi, Jamnik, Mateja, Liò, Pietro

arXiv.org Artificial Intelligence

The success of Large Language Models (LLMs), e.g., ChatGPT, is witnessed by their planetary popularity, their capability of human-like communication, and also by their steadily improved reasoning performance. However, it remains unclear whether LLMs reason. It is an open problem how traditional neural networks can be qualitatively extended to go beyond the statistic paradigm and achieve high-level cognition. Here, we present a novel qualitative extension by generalising computational building blocks from vectors to spheres. We propose Sphere Neural Networks (SphNNs) for human-like reasoning through model construction and inspection, and develop SphNN for syllogistic reasoning, a microcosm of human rationality. SphNN is a hierarchical neuro-symbolic Kolmogorov-Arnold geometric GNN, and uses a neuro-symbolic transition map of neighbourhood spatial relations to transform the current sphere configuration towards the target. SphNN is the first neural model that can determine the validity of long-chained syllogistic reasoning in one epoch without training data, with the worst computational complexity of O(N). SphNN can evolve into various types of reasoning, such as spatio-temporal reasoning, logical reasoning with negation and disjunction, event reasoning, neuro-symbolic unification, and humour understanding (the highest level of cognition). All these suggest a new kind of Herbert A. Simon's scissors with two neural blades. SphNNs will tremendously enhance interdisciplinary collaborations to develop the two neural blades and realise deterministic neural reasoning and human-bounded rationality and elevate LLMs to reliable psychological AI. This work suggests that the non-zero radii of spheres are the missing components that prevent traditional deep-learning systems from reaching the realm of rational reasoning and cause LLMs to be trapped in the swamp of hallucination.


Autism, dyslexia, ADHD. How the University of San Diego is helping 'neurodivergent' students succeed

Los Angeles Times

University of San Diego professors are developing programs to empower neurodivergent students --- those with autism spectrum disorder, ADHD, dyslexia, among other learning differences.


Anomaly Detection for Incident Response at Scale

Wang, Hanzhang, Tangirala, Gowtham Kumar, Naidu, Gilkara Pranav, Mayville, Charles, Roy, Arighna, Sun, Joanne, Mandava, Ramesh Babu

arXiv.org Artificial Intelligence

We present a machine learning-based anomaly detection product, AI Detect and Respond (AIDR), that monitors Walmart's business and system health in real-time. During the validation over 3 months, the product served predictions from over 3000 models to more than 25 application, platform, and operation teams, covering 63\% of major incidents and reducing the mean-time-to-detect (MTTD) by more than 7 minutes. Unlike previous anomaly detection methods, our solution leverages statistical, ML and deep learning models while continuing to incorporate rule-based static thresholds to incorporate domain-specific knowledge. Both univariate and multivariate ML models are deployed and maintained through distributed services for scalability and high availability. AIDR has a feedback loop that assesses model quality with a combination of drift detection algorithms and customer feedback. It also offers self-onboarding capabilities and customizability. AIDR has achieved success with various internal teams with lower time to detection and fewer false positives than previous methods. As we move forward, we aim to expand incident coverage and prevention, reduce noise, and integrate further with root cause recommendation (RCR) to enable an end-to-end AIDR experience.


Hierarchical Attention Models for Multi-Relational Graphs

Iyer, Roshni G., Wang, Wei, Sun, Yizhou

arXiv.org Artificial Intelligence

We present Bi-Level Attention-Based Relational Graph Convolutional Networks (BR-GCN), unique neural network architectures that utilize masked self-attentional layers with relational graph convolutions, to effectively operate on highly multi-relational data. BR-GCN models use bi-level attention to learn node embeddings through (1) node-level attention, and (2) relation-level attention. The node-level self-attentional layers use intra-relational graph interactions to learn relation-specific node embeddings using a weighted aggregation of neighborhood features in a sparse subgraph region. The relation-level self-attentional layers use inter-relational graph interactions to learn the final node embeddings using a weighted aggregation of relation-specific node embeddings. The BR-GCN bi-level attention mechanism extends Transformer-based multiplicative attention from the natural language processing (NLP) domain, and Graph Attention Networks (GAT)-based attention, to large-scale heterogeneous graphs (HGs). On node classification, BR-GCN outperforms baselines from 0.29% to 14.95% as a stand-alone model, and on link prediction, BR-GCN outperforms baselines from 0.02% to 7.40% as an auto-encoder model. We also conduct ablation studies to evaluate the quality of BR-GCN's relation-level attention and discuss how its learning of graph structure may be transferred to enrich other graph neural networks (GNNs). Through various experiments, we show that BR-GCN's attention mechanism is both scalable and more effective in learning compared to state-of-the-art GNNs.


As California fires worsen, can AI come to the rescue?

Los Angeles Times

Just before 3 a.m. one night this month, Scott Slumpff was awakened by the ding of a text message. "An ALERTCalifornia anomaly has been confirmed in your area of interest," the message said. Slumpff, a battalion chief with the California Department of Forestry and Fire Protection, sprang into action. The message meant the agency's new artificial intelligence system had identified signs of a wildfire with a remote mountaintop camera in San Diego County. Within minutes, crews were dispatched to the burgeoning blaze on Mount Laguna -- squelching it before it grew any larger than a 10-foot-by-10-foot spot.