Goto

Collaborating Authors

 interacting


RoleAgent: Building, Interacting, and Benchmarking High-quality Role-Playing Agents from Scripts

Neural Information Processing Systems

Believable agents can empower interactive applications ranging from immersive environments to rehearsal spaces for interpersonal communication. Recently, generative agents have been proposed to simulate believable human behavior by using Large Language Models. However, the existing method heavily relies on human-annotated agent profiles (e.g., name, age, personality, relationships with others, and so on) for the initialization of each agent, which cannot be scaled up easily. In this paper, we propose a scalable RoleAgent framework to generate high-quality role-playing agents from raw scripts, which includes building and interacting stages. Specifically, in the building stage, we use a hierarchical memory system to extract and summarize the structure and high-level information of each agent for the raw script.


Interacting with AI Reasoning Models: Harnessing "Thoughts" for AI-Driven Software Engineering

Treude, Christoph, Kula, Raula Gaikovina

arXiv.org Artificial Intelligence

Recent advances in AI reasoning models provide unprecedented transparency into their decision-making processes, transforming them from traditional black-box systems into models that articulate step-by-step chains of thought rather than producing opaque outputs. This shift has the potential to improve software quality, explainability, and trust in AI-augmented development. However, software engineers rarely have the time or cognitive bandwidth to analyze, verify, and interpret every AI-generated thought in detail. Without an effective interface, this transparency could become a burden rather than a benefit. In this paper, we propose a vision for structuring the interaction between AI reasoning models and software engineers to maximize trust, efficiency, and decision-making power. We argue that simply exposing AI's reasoning is not enough -- software engineers need tools and frameworks that selectively highlight critical insights, filter out noise, and facilitate rapid validation of key assumptions. To illustrate this challenge, we present motivating examples in which AI reasoning models state their assumptions when deciding which external library to use and produce divergent reasoning paths and recommendations about security vulnerabilities, highlighting the need for an interface that prioritizes actionable insights while managing uncertainty and resolving conflicts. We then outline a research roadmap for integrating automated summarization, assumption validation, and multi-model conflict resolution into software engineering workflows. Achieving this vision will unlock the full potential of AI reasoning models to enable software engineers to make faster, more informed decisions without being overwhelmed by unnecessary detail.


A Paragraph is All It Takes: Rich Robot Behaviors from Interacting, Trusted LLMs

OpenMind, null, Zhong, Shaohong, Zhou, Adam, Chen, Boyuan, Luo, Homin, Liphardt, Jan

arXiv.org Artificial Intelligence

Large Language Models (LLMs) are compact representations of all public knowledge of our physical environment and animal and human behaviors. The application of LLMs to robotics may offer a path to highly capable robots that perform well across most human tasks with limited or even zero tuning. Aside from increasingly sophisticated reasoning and task planning, networks of (suitably designed) LLMs offer ease of upgrading capabilities and allow humans to directly observe the robot's thinking. Here we explore the advantages, limitations, and particularities of using LLMs to control physical robots. The basic system consists of four LLMs communicating via a human language data bus implemented via web sockets and ROS2 message passing. Surprisingly, rich robot behaviors and good performance across different tasks could be achieved despite the robot's data fusion cycle running at only 1Hz and the central data bus running at the extremely limited rates of the human brain, of around 40 bits/s. The use of natural language for inter-LLM communication allowed the robot's reasoning and decision making to be directly observed by humans and made it trivial to bias the system's behavior with sets of rules written in plain English. These rules were immutably written into Ethereum, a global, public, and censorship resistant Turing-complete computer. We suggest that by using natural language as the data bus among interacting AIs, and immutable public ledgers to store behavior constraints, it is possible to build robots that combine unexpectedly rich performance, upgradability, and durable alignment with humans.


The Conversation is the Command: Interacting with Real-World Autonomous Robot Through Natural Language

Nwankwo, Linus, Rueckert, Elmar

arXiv.org Artificial Intelligence

In recent years, autonomous agents have surged in real-world environments such as our homes, offices, and public spaces. However, natural human-robot interaction remains a key challenge. In this paper, we introduce an approach that synergistically exploits the capabilities of large language models (LLMs) and multimodal vision-language models (VLMs) to enable humans to interact naturally with autonomous robots through conversational dialogue. We leveraged the LLMs to decode the high-level natural language instructions from humans and abstract them into precise robot actionable commands or queries. Further, we utilised the VLMs to provide a visual and semantic understanding of the robot's task environment. Our results with 99.13% command recognition accuracy and 97.96% commands execution success show that our approach can enhance human-robot interaction in real-world applications. The video demonstrations of this paper can be found at https://osf.io/wzyf6 and the code is available at our GitHub repository (https://github.com/LinusNEP/TCC_IRoNL.git).


AIxArtist: A First-Person Tale of Interacting with Artificial Intelligence to Escape Creative Block

Lewis, Makayla

arXiv.org Artificial Intelligence

The future of the arts and artificial intelligence (AI) is promising as technology advances. As the use of AI in design becomes more widespread, art practice may not be a human-only art form and could instead become a digitally integrated experience. With enhanced creativity and collaboration, arts and AI could work together towards creating artistic outputs that are visually appealing and meet the needs of the artist and viewer. While it is uncertain how far the integration will go, arts and AI will likely influence one another. This workshop pictorial puts forward first-person research that shares interactions between an HCI researcher and AI as they try to escape the creative block. The pictorial paper explores two questions: How can AI support artists' creativity, and what does it mean to be explainable in this context? HIs, ChatGPT and Midjourney were engaged; the result was a series of reflections that require further discussion and explorations in the XAIxArts community: Transparency of attribution, the creation process, ethics of asking, and inspiration vs copying.


Interacting with next-phrase suggestions: How suggestion systems aid and influence the cognitive processes of writing

Bhat, Advait, Agashe, Saaket, Mohile, Niharika, Oberoi, Parth, Jangir, Ravi, Joshi, Anirudha

arXiv.org Artificial Intelligence

Writing with next-phrase suggestions powered by large language models is becoming more pervasive by the day. However, research to understand writers' interaction and decision-making processes while engaging with such systems is still emerging. We conducted a qualitative study to shed light on writers' cognitive processes while writing with next-phrase suggestion systems. To do so, we recruited 14 amateur writers to write two reviews each, one without suggestions and one with suggestions. Additionally, we also positively and negatively biased the suggestion system to get a diverse range of instances where writers' opinions and the bias in the language model align or misalign to varying degrees. We found that writers interact with next-phrase suggestions in various complex ways: Writers abstracted and extracted multiple parts of the suggestions and incorporated them within their writing, even when they disagreed with the suggestion as a whole; along with evaluating the suggestions on various criteria. The suggestion system also had various effects on the writing process, such as altering the writer's usual writing plans, leading to higher levels of distraction etc. Based on our qualitative analysis using the cognitive process model of writing by Hayes as a lens, we propose a theoretical model of 'writer-suggestion interaction' for writing with GPT-2 (and causal language models in general) for a movie review writing task, followed by directions for future research and design.


LAIDEN - Interacting with Robots and AI

#artificialintelligence

The perception of agency and the way we explain an agent behavior play a role in the interaction with agentic things, like robotic object and intelligent playthings. Children easily imbue robotic objects and playthings with agency and explain their behavior on psychological term attributing intelligence relying on agency perception. However, human robot interaction researchers and practitioners often do not tackle in the design process how children make sense of a robot, its agency and how children explain a robot's behavior and intelligence. In this talk, I will shed light on these phenomena and their implications for child-interactions. I will argue that responsible human-centred design should play a more prominent role in the human-robot interaction design cycle to understand how children perceive a robot's agency and explain a robot's behavior.


Learning to Prove Theorems via Interacting with Proof Assistants

Yang, Kaiyu, Deng, Jia

arXiv.org Machine Learning

Humans prove theorems by relying on substantial high-level reasoning and problem-specific insights. Proof assistants offer a formalism that resembles human mathematical reasoning, representing theorems in higher-order logic and proofs as high-level tactics. However, human experts have to construct proofs manually by entering tactics into the proof assistant. In this paper, we study the problem of using machine learning to automate the interaction with proof assistants. We construct CoqGym, a large-scale dataset and learning environment containing 71K human-written proofs from 123 projects developed with the Coq proof assistant. We develop ASTactic, a deep learning-based model that generates tactics as programs in the form of abstract syntax trees (ASTs). Experiments show that ASTactic trained on CoqGym can generate effective tactics and can be used to prove new theorems not previously provable by automated methods. Code is available at https://github.com/princeton-vl/CoqGym.


Interacting with this therapy bot can help children with autism perfect their social skills

#artificialintelligence

A child with autism spectrum disorder (ASD) might have trouble communicating verbally, paying attention to others, or controlling their stress and anxiety. These difficulties can affect the child's social life and their success in school. Now, a team of researchers from robotics startup LuxAI have created QTrobot, a bot designed to help children with autism learn valuable social skills. They plan to present the results of a QTrobot study at RO-MAN 2018, a symposium on robot and human interactive communication, on August 28. QTrobot is just over two feet tall, with a humanoid body and a screen where a person's face would be.


What Interacting With Robots Might Reveal About Human Nature

The Atlantic - Technology

Robot panic seems to move in cycles, as new innovations in technology drive fear about machines that will take over our jobs, our lives, and our society--only to collapse as it becomes clear just how far away such omnipotent robots are. Today's robots can barely walk effectively, much less conquer civilization. But that doesn't mean there aren't good reasons to be nervous. The more pressing problem today is not what robots can do our bodies and livelihoods, but what they will do to our brains. "The problem is not that if we teach robots to kick they'll kick our ass," Kate Darling, an MIT robot ethicist, said Thursday at the Aspen Ideas Festival, which is co-hosted by the Aspen Institute and The Atlantic.