Goto

Collaborating Authors

 grandma


'Jury Duty Presents: Company Retreat' Almost Makes Corporate Culture Seem Fun

WIRED

The Amazon Prime prank series amplifies the hijinks of workplace dynamics, while showing how people find purpose--and community--in their jobs despite impossible situations. Anthony Norman is your typical Gen Z worker: 25, a little wayward, and struggling to find a full time job. Unemployment rates are high . AI is creating a crisis for young people trying to enter the workforce. And several companies--including Amazon, Block, and Meta --have embraced tech's latest era of layoffmaxxing, with some cutting their staff by 20 percent.


ALAS: Transactional and Dynamic Multi-Agent LLM Planning

Geng, Longling, Chang, Edward Y.

arXiv.org Artificial Intelligence

Large language models enable flexible multi-agent planning but remain fragile in practice: verification is often circular, state changes are not tracked for repair, and small faults trigger costly global recomputation. We present ALAS, a stateful, disruption-aware framework that separates planning from non-circular validation, records a versioned execution log for grounded checks and restore points, and performs localized repair that preserves work in progress. The validator operates independently of the planning LLM with fresh, bounded context, avoiding self-check loops and mid-context attrition. The repair protocol edits only the minimal affected region under explicit policies (retry, catch, timeout, backoff, idempotency keys, compensation, loop guards) defined in a canonical workflow IR that maps to Amazon States Language and Argo Workflows. On job-shop scheduling suites (DMU, TA) across five classical benchmarks, ALAS matches or exceeds strong single-LLM and multi-agent baselines, achieving 83.7% success, reducing token usage by 60%, and running 1.82times faster under comparable settings. A minimal reliability study shows that the validator detects injected structural faults with low overhead, and that localized repair contains runtime perturbations with a bounded edit radius and less makespan degradation than global recompute. Results indicate that the combination of validator isolation, versioned execution logs, and localized repair provides measurable efficiency, feasibility, and scalability for multi-agent LLM planning. Code and seeds will be released.


Solve it with EASE

Viktorin, Adam, Kadavy, Tomas, Kovac, Jozef, Pluhacek, Michal, Senkerik, Roman

arXiv.org Artificial Intelligence

This paper presents EASE (Effortless Algorithmic Solution Evolution), an open-source and fully modular framework for iterative algorithmic solution generation leveraging large language models (LLMs). EASE integrates generation, testing, analysis, and evaluation into a reproducible feedback loop, giving users full control over error handling, analysis, and quality assessment. Its architecture supports the orchestration of multiple LLMs in complementary roles-such as generator, analyst, and evaluator. By abstracting the complexity of prompt design and model management, EASE provides a transparent and extensible platform for researchers and practitioners to co-design algorithms and other generative solutions across diverse domains.


Interview with Kate Candon: Leveraging explicit and implicit feedback in human-robot interactions

AIHub

In this interview series, we're meeting some of the AAAI/SIGAI Doctoral Consortium participants to find out more about their research. Kate Candon is a PhD student at Yale University interested in understanding how we can create interactive agents that are more effectively able to help people. We spoke to Kate to find out more about how she is leveraging explicit and implicit feedback in human-robot interactions. Specifically I'm interested in how we can get robots to better learn from humans in the way that they naturally teach. Typically, a lot of work in robot learning is with a human teacher who is only tasked with giving explicit feedback to the robot, but they're not necessarily engaged in the task.


ALAS: A Stateful Multi-LLM Agent Framework for Disruption-Aware Planning

Chang, Edward Y., Geng, Longling

arXiv.org Artificial Intelligence

Large language models (LLMs) excel at rapid generation of text and multimodal content, yet they falter on transaction-style planning that demands ACID-like guarantees and real-time disruption recovery. We present Adaptive LLM Agent System (ALAS), a framework that tackles four fundamental LLM deficits: (i) absence of self-verification, (ii) context erosion, (iii) next-token myopia, and (iv) lack of persistent state. ALAS decomposes each plan into role-specialized agents, equips them with automatic state tracking, and coordinates them through a lightweight protocol. When disruptions arise, agents apply history-aware local compensation, avoiding costly global replanning and containing cascade effects. On real-world, large-scale job-shop scheduling benchmarks, ALAS sets new best results for static sequential planning and excels in dynamic reactive scenarios with unexpected disruptions. These gains show that principled modularization plus targeted compensation can unlock scalable and resilient planning with LLMs.


Addressing Moral Uncertainty using Large Language Models for Ethical Decision-Making

Dubey, Rohit K., Dailisan, Damian, Mahajan, Sachit

arXiv.org Artificial Intelligence

We present an ethical decision-making framework that refines a pre-trained reinforcement learning (RL) model using a task-agnostic ethical layer. Following initial training, the RL model undergoes ethical fine-tuning, where human feedback is replaced by feedback generated from a large language model (LLM). The LLM embodies consequentialist, deontological, virtue, social justice, and care ethics as moral principles to assign belief values to recommended actions during ethical decision-making. An ethical layer aggregates belief scores from multiple LLM-derived moral perspectives using Belief Jensen-Shannon Divergence and Dempster-Shafer Theory into probability scores that also serve as the shaping reward, steering the agent toward choices that align with a balanced ethical framework. This integrated learning framework helps the RL agent navigate moral uncertainty in complex environments and enables it to make morally sound decisions across diverse tasks. Our approach, tested across different LLM variants and compared with other belief aggregation techniques, demonstrates improved consistency, adaptability, and reduced reliance on handcrafted ethical rewards. This method is especially effective in dynamic scenarios where ethical challenges arise unexpectedly, making it well-suited for real-world applications.


Fox News AI Newsletter: AI app helps you turn anything into LEGO models

FOX News

BUILD LEGO CREATIONS: This innovative app is here to make custom Lego creation fun and accessible for everyone, whether you're a seasoned builder or just getting started. By using advanced artificial intelligence and mobile scanning technology, Brick My World opens up a world of creative possibilities. 'OUR HOLIDAY GIFT': OpenAI released its text-to-video artificial intelligence model, Sora, this week after the completion of its testing phase. The OpenAI logo is being displayed on a smartphone with the Sora text-to-video generator visible in the background in this photo illustration, taken in Brussels, Belgium, on February 16, 2024. GRANNY FIGHTS BACK: Daisy is an artificial intelligence-powered grandma developed by Virgin Media O2 to interact with scammers.


The AI-powered grandma taking on scammers

FOX News

Daisy is an artificial intelligence-powered grandma created to interact with scammers. Are you tired of scammers calling your phone, trying to trick you into giving away your hard-earned money? Many people are fed up with the constant barrage of fraudulent calls and messages. But what if you could fight back in a fun and creative way? Enter the world of scambaiting, where people waste scammers' time and resources instead of falling for their tricks.


A recipe for magical realism: Gabriel García Márquez and a video game about potatoes

The Guardian

Sopa (the Spanish for "soup") is a game about a young boy who goes to fetch a potato for his grandma, then stumbles upon a magical world at the back of the food cupboard. "The pantry seems to get longer and longer," explains creative director Juan Castañeda. "And when you're about to grab the sack of potatoes, you get pulled into this other world of fantasy and magical realism. So you go on all these adventures, and meet all these different characters, but at the end of the day, you're really just trying to get that potato for your grandma's soup." As video game quests go, this is fabulously mundane and makes a refreshing change from rescuing princesses in castles and saving lands in peril.


Hidding the Ghostwriters: An Adversarial Evaluation of AI-Generated Student Essay Detection

Peng, Xinlin, Zhou, Ying, He, Ben, Sun, Le, Sun, Yingfei

arXiv.org Artificial Intelligence

Large language models (LLMs) have exhibited remarkable capabilities in text generation tasks. However, the utilization of these models carries inherent risks, including but not limited to plagiarism, the dissemination of fake news, and issues in educational exercises. Although several detectors have been proposed to address these concerns, their effectiveness against adversarial perturbations, specifically in the context of student essay writing, remains largely unexplored. This paper aims to bridge this gap by constructing AIG-ASAP, an AI-generated student essay dataset, employing a range of text perturbation methods that are expected to generate high-quality essays while evading detection. Through empirical experiments, we assess the performance of current AIGC detectors on the AIG-ASAP dataset. The results reveal that the existing detectors can be easily circumvented using straightforward automatic adversarial attacks. Specifically, we explore word substitution and sentence substitution perturbation methods that effectively evade detection while maintaining the quality of the generated essays. This highlights the urgent need for more accurate and robust methods to detect AI-generated student essays in the education domain.