aria
New rules confirm public has a right to see how UK government uses AI
Government departments and other public bodies in the UK must consider requests to release information about AI-produced content, regulators have confirmed. The move follows a successful request by New Scientist for the release of a minister's ChatGPT logs The use of AI chatbots is subject to the UK's Freedom of Information laws Text, images and other content produced by UK government departments and other public bodies using artificial intelligence are subject to freedom of information (FOI) laws, regulators have confirmed - potentially opening the door for the public to gain access to ministers' ChatGPT or other chatbot records. The Information Commissioner's Office (ICO), the UK's data-protection agency, has released new guidance confirming that "If staff at a public authority use AI for work purposes, the information generated will be subject to FOIA [the Freedom of Information Act] along with the prompts used". Last year, successfully requested the then-UK tech secretary Peter Kyle's ChatGPT logs under FOI legislation, in what is believed to be a world first. That triggered subsequent requests from other news outlets to obtain other information, but many have either been rejected on cost grounds or labelled as "vexatious", an umbrella term that allows authorities to reject a request.
UK 'invention agency' grants 50m of public money to US tech and venture capital firms
OpenAI's Sam Altman, left, is a backer of Rain Neuromophics, one of the companies that received funds from the UK's Aria, the brainchild of Dominic Cummings, right OpenAI's Sam Altman, left, is a backer of Rain Neuromophics, one of the companies that received funds from the UK's Aria, the brainchild of Dominic Cummings, right Exclusive: Brainchild of Dominic Cummings, Aria is aimed at funding'crazy' scientific projects to benefit the UK Britain's "invention agency" has pledged £50m of UK taxpayer money to US tech companies and venture capital projects. Dreamed up by Dominic Cummings to fund "crazy" ideas, the Advanced Research and Invention Agency (Aria) is meant to " restore Britain's place as a scientific superpower ". But a joint investigation by the Guardian and Democracy for Sale, an investigative website, has established that more than an eighth of the agency's £400m in research and development funding over the past two years has gone to 14 US tech companies and venture capital groups, in some cases, with no clear return for the UK or Aria. One of these companies, Rain Neuromorphics, is also backed by the OpenAI chief executive, Sam Altman, and was reported to be near collapse last year, shortly after winning Aria money. It did not respond to a request for comment; two of its founders appear to have left the company.
- Government (1.00)
- Information Technology (0.90)
- Banking & Finance > Capital Markets (0.83)
- Leisure & Entertainment > Sports (0.70)
- Information Technology > Communications > Social Media (0.72)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (0.66)
- Information Technology > Artificial Intelligence > Natural Language > Chatbot (0.66)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning > Generative AI (0.66)
The UK's Answer to Darpa Wants to Rewire the Human Brain
ARIA has a billion-dollar budget and big aspirations for tackling everything from epilepsy to Alzheimer's. The UK's Advanced Research and Innovation Agency (ARIA) was established in 2023 with the goal of pursuing "high-risk, high-reward" moonshots in sectors ranging from bolstering food security to new ways of ramping up human immunity . With more than £1 billion (about $1.3 billion) worth of government funding earmarked between now and 2030, one of ARIA's most ambitious programs is a £69 million initiative that aims to develop more tailored ways of modulating the human brain. The hope is to eventually address an entire range of disorders, from epilepsy to Alzheimer's. Reports have previously estimated that this suite of neurological conditions costs the UK economy tens of billions of dollars each year.
- Europe (0.69)
- North America > United States > California (0.15)
The UK government is backing AI that can run its own lab experiments
A competition calling for research projects involving so-called AI scientists shows just how fast this technology is moving. A number of startups and universities that are building "AI scientists" to design and run experiments in the lab, including robot biologists and chemists, have just won extra funding from the UK government agency that funds moonshot R&D. The competition, set up by ARIA (the Advanced Research and Invention Agency), gives a clear sense of how fast this technology is moving: The agency received 245 proposals from research teams that are already building tools capable of automating increasing amounts of lab work. ARIA defines an AI scientist as a system that can run an entire scientific workflow, coming up with hypotheses, designing and running experiments to test those hypotheses, and then analyzing the results. In many cases, the system may then feed those results back into itself and run the loop again and again. Human scientists become overseers, coming up with the initial research questions and then letting the AI scientist get on with the grunt work.
- North America > United States > Massachusetts (0.05)
- Europe > United Kingdom > England > Merseyside > Liverpool (0.05)
- Asia > India (0.05)
- Asia > China (0.05)
The Ghost in the Keys: A Disklavier Demo for Human-AI Musical Co-Creativity
Bradshaw, Louis, Spangher, Alexander, Biderman, Stella, Colton, Simon
While generative models for music composition are increasingly capable, their adoption by musicians is hindered by text-prompting, an asynchronous workflow disconnected from the embodied, responsive nature of instrumental performance. To address this, we introduce Aria-Duet, an interactive system facilitating a real-time musical duet between a human pianist and Aria, a state-of-the-art generative model, using a Yamaha Disklavier as a shared physical interface. The framework enables a turn-taking collaboration: the user performs, signals a handover, and the model generates a coherent continuation performed acoustically on the piano. Beyond describing the technical architecture enabling this low-latency interaction, we analyze the system's output from a musicological perspective, finding the model can maintain stylistic semantics and develop coherent phrasal ideas, demonstrating that such embodied systems can engage in musically sophisticated dialogue and open a promising new path for human-AI co-creation.
- North America > United States > Illinois > Cook County > Chicago (0.04)
- North America > United States > California > Alameda County > Berkeley (0.04)
- Europe > United Kingdom > England > Oxfordshire > Oxford (0.04)
- (2 more...)
- Media > Music (1.00)
- Leisure & Entertainment (1.00)
Spec-Driven AI for Science: The ARIA Framework for Automated and Reproducible Data Analysis
Chen, Chuke, Luo, Biao, Li, Nan, Wang, Boxiang, Yang, Hang, Guo, Jing, Xu, Ming
The rapid expansion of scientific data has widened the gap between analytical capability and research intent. Existing AI-based analysis tools, ranging from AutoML frameworks to agentic research assistants, either favor automation over transparency or depend on manual scripting that hinders scalability and reproducibility. We present ARIA (Automated Research Intelligence Assistant), a spec-driven, human-in-the-loop framework for automated and interpretable data analysis. ARIA integrates six interoperable layers, namely Command, Context, Code, Data, Orchestration, and AI Module, within a document-centric workflow that unifies human reasoning and machine execution. Through natural-language specifications, researchers define analytical goals while ARIA autonomously generates executable code, validates computations, and produces transparent documentation. Beyond achieving high predictive accuracy, ARIA can rapidly identify optimal feature sets and select suitable models, minimizing redundant tuning and repetitive experimentation. In the Boston Housing case, ARIA discovered 25 key features and determined XGBoost as the best performing model (R square = 0.93) with minimal overfitting. Evaluations across heterogeneous domains demonstrate ARIA's strong performance, interpretability, and efficiency compared with state-of-the-art systems. By combining AI for research and AI for science principles within a spec-driven architecture, ARIA establishes a new paradigm for transparent, collaborative, and reproducible scientific discovery.
- Asia > China > Beijing > Beijing (0.05)
- Asia > China > Shanghai > Shanghai (0.04)
- North America > United States (0.04)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Agents (0.70)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Scientific Discovery (0.69)
- Information Technology > Artificial Intelligence > Machine Learning > Statistical Learning (0.49)
- (2 more...)
Enabling Self-Improving Agents to Learn at Test Time With Human-In-The-Loop Guidance
He, Yufei, Li, Ruoyu, Chen, Alex, Liu, Yue, Chen, Yulin, Sui, Yuan, Chen, Cheng, Zhu, Yi, Luo, Luca, Yang, Frank, Hooi, Bryan
Large language model (LLM) agents often struggle in environments where rules and required domain knowledge frequently change, such as regulatory compliance and user risk screening. Current approaches, like offline fine-tuning and standard prompting, are insufficient because they cannot effectively adapt to new knowledge during actual operation. To address this limitation, we propose the Adaptive Reflective Interactive Agent (ARIA), an LLM agent framework designed specifically to continuously learn updated domain knowledge at test time. ARIA assesses its own uncertainty through structured self-dialogue, proactively identifying knowledge gaps and requesting targeted explanations or corrections from human experts. It then systematically updates an internal, timestamped knowledge repository with provided human guidance, detecting and resolving conflicting or outdated knowledge through comparisons and clarification queries. We evaluate ARIA on the realistic customer due diligence name screening task on TikTok Pay, alongside publicly available dynamic knowledge tasks. Results demonstrate significant improvements in adaptability and accuracy compared to baselines using standard offline fine-tuning and existing self-improving agents. ARIA is deployed within TikTok Pay serving over 150 million monthly active users, confirming its practicality and effectiveness for operational use in rapidly evolving environments.
- Law (1.00)
- Information Technology > Security & Privacy (1.00)
- Information Technology > Communications > Social Media (1.00)
- Information Technology > Artificial Intelligence > Representation & Reasoning (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.68)
Benchmarking Egocentric Visual-Inertial SLAM at City Scale
Krishnan, Anusha, Liu, Shaohui, Sarlin, Paul-Edouard, Gentilhomme, Oscar, Caruso, David, Monge, Maurizio, Newcombe, Richard, Engel, Jakob, Pollefeys, Marc
Precise 6-DoF simultaneous localization and mapping (SLAM) from onboard sensors is critical for wearable devices capturing egocentric data, which exhibits specific challenges, such as a wider diversity of motions and viewpoints, prevalent dynamic visual content, or long sessions affected by time-varying sensor calibration. While recent progress on SLAM has been swift, academic research is still driven by benchmarks that do not reflect these challenges or do not offer sufficiently accurate ground truth poses. In this paper, we introduce a new dataset and benchmark for visual-inertial SLAM with egocentric, multi-modal data. We record hours and kilometers of trajectories through a city center with glasses-like devices equipped with various sensors. We leverage surveying tools to obtain control points as indirect pose annotations that are metric, centimeter-accurate, and available at city scale. This makes it possible to evaluate extreme trajectories that involve walking at night or traveling in a vehicle. We show that state-of-the-art systems developed by academia are not robust to these challenges and we identify components that are responsible for this. In addition, we design tracks with different levels of difficulty to ease in-depth analysis and evaluation of less mature approaches. The dataset and benchmark are available at https://www.lamaria.ethz.ch.
- Europe > Switzerland > Zürich > Zürich (0.04)
- South America > Chile > Santiago Metropolitan Region > Santiago Province > Santiago (0.04)
- Asia > Japan > Honshū > Chūbu > Ishikawa Prefecture > Kanazawa (0.04)
Sexy robot unveiled at Vegas tech conference is world's most realistic
You could be looking at the ultra-realistic and creepy future of sex dolls. A new'companion robot' made its debut at a major tech event in Vegas this week, sporting perfectly shaped breasts, a pert buttocks, thick lips and features of a youthful 20-something. Its creators, Realbotix, claim the 175,000 bot is to keep elderly gentlemen from being lonely. Asked why it had been shaped in an appealing way, the creators said it was because men wanted something nice to look at. Dressed in a black tracksuit, when asking the robot questions it gives long responses while having slightly jerky hand and body movements.
HOT3D: Hand and Object Tracking in 3D from Egocentric Multi-View Videos
Banerjee, Prithviraj, Shkodrani, Sindi, Moulon, Pierre, Hampali, Shreyas, Han, Shangchen, Zhang, Fan, Zhang, Linguang, Fountain, Jade, Miller, Edward, Basol, Selen, Newcombe, Richard, Wang, Robert, Engel, Jakob Julian, Hodan, Tomas
We introduce HOT3D, a publicly available dataset for egocentric hand and object tracking in 3D. The dataset offers over 833 minutes (more than 3.7M images) of multi-view RGB/monochrome image streams showing 19 subjects interacting with 33 diverse rigid objects, multi-modal signals such as eye gaze or scene point clouds, as well as comprehensive ground-truth annotations including 3D poses of objects, hands, and cameras, and 3D models of hands and objects. In addition to simple pick-up/observe/put-down actions, HOT3D contains scenarios resembling typical actions in a kitchen, office, and living room environment. The dataset is recorded by two head-mounted devices from Meta: Project Aria, a research prototype of light-weight AR/AI glasses, and Quest 3, a production VR headset sold in millions of units. Ground-truth poses were obtained by a professional motion-capture system using small optical markers attached to hands and objects. Hand annotations are provided in the UmeTrack and MANO formats and objects are represented by 3D meshes with PBR materials obtained by an in-house scanner. In our experiments, we demonstrate the effectiveness of multi-view egocentric data for three popular tasks: 3D hand tracking, 6DoF object pose estimation, and 3D lifting of unknown in-hand objects. The evaluated multi-view methods, whose benchmarking is uniquely enabled by HOT3D, significantly outperform their single-view counterparts.
- North America > United States > Massachusetts (0.04)
- Asia > Middle East > Saudi Arabia > Asir Province > Abha (0.04)
- Asia > Japan > Honshū > Chūbu > Ishikawa Prefecture > Kanazawa (0.04)