Goto

Collaborating Authors

 ai software


AgentSLA : Towards a Service Level Agreement for AI Agents

Jouneaux, Gwendal, Cabot, Jordi

arXiv.org Artificial Intelligence

AI components are increasingly becoming a key element of all types of software systems to enhance their functionality. These AI components are often implemented as AI Agents, offering more autonomy than a plain integration of Large Language Models (LLMs), moving from a Model-as-a-Service paradigm to an Agent-as-a-Service one, bringing new challenges to the development of smart software systems. Indeed, while support for the design, implementation, and deployment of those agents exist, the specification of Quality of Service (QoS) and definition of Service Level Agreements (SLAs) aspects for those agents, important to ensure the quality of the resulting systems, remains an open challenge. Part of this is due to the difficulty to clearly define quality in the context of AI components, resulting in a lack of consensus on how to best approach Quality Assurance (QA) for these types of systems. To address this challenge, this paper proposes both a quality model for AI agents based on the ISO/IEC 25010 standard, and a domain specific language to support the definition of SLAs for the services provided by these AI agents.


The Vibes-Based Pricing of 'Pro' AI Software

WIRED

Hosts Lauren Goode and and Michael Calore speak with staff writer Reece Rogers to find out what's behind these models that AI companies bill as their most powerful, and whether they could become a staple in our future. Write to us at uncannyvalley@wired.com. You can always listen to this week's podcast through the audio player on this page, but if you want to subscribe for free to get every episode, here's how: If you're on an iPhone or iPad, open the app called Podcasts, or just tap this link. Note: This is an automated transcript, which may contain errors. Michael Calore: Hey, this is Mike.


OpenAI releases two 'open' AI models after DeepSeek's success

The Japan Times

OpenAI is releasing a pair of open and freely available artificial intelligence models that can mimic the human process of reasoning, months after China's DeepSeek gained global attention with its own open AI software. The two models, called GPT-oss-120b and GPT-oss-20b, will be available on AI software hosting platform Hugging Face and can produce text -- but not images or videos -- in response to user prompts, OpenAI said on Tuesday. These models can also carry out complex tasks like writing code and looking up information online on a user's behalf, the company said. Crucially, the models are both open-weight systems, similar to Meta Platforms' Llama. The term "weight" refers to the parameters in an AI model.


A Hiker Was Missing for Nearly a Year--Until an AI System Recognized His Helmet

WIRED

How long does it take to identify the helmet of a hiker lost in a 183-hectare mountain area, analyzing 2,600 frames taken by a drone from approximately 50 meters away? If done with a human eye, weeks or months. If analyzed by an artificial intelligence system, one afternoon. The National Alpine and Speleological Rescue Corps, known by it's Italian initialism CNSAS, relied on AI to find the body of a person missing in Italy's Piedmont region on the north face of Monviso--the highest peak in the Cottian Alps--since September 2024. According to Saverio Isola, the CNSAS drone pilot who intervened along with his colleague Giorgio Viana, the operation--including searching for any sign of the missing hiker, the discovery and recovery of his body, and a stoppage due to bad weather--lasted less than three days.


Multi-Centre Validation of a Deep Learning Model for Scoliosis Assessment

Kubov, Šimon, Klíčník, Simon, Dandár, Jakub, Straka, Zdeněk, Kvaková, Karolína, Kvak, Daniel

arXiv.org Artificial Intelligence

Scoliosis affects roughly 2 to 4 percent of adolescents, and treatment decisions depend on precise Cobb angle measurement. Manual assessment is time consuming and subject to inter observer variation. We conducted a retrospective, multi centre evaluation of a fully automated deep learning software (Carebot AI Bones, Spine Measurement functionality; Carebot s.r.o.) on 103 standing anteroposterior whole spine radiographs collected from ten hospitals. Two musculoskeletal radiologists independently measured each study and served as reference readers. Agreement between the AI and each radiologist was assessed with Bland Altman analysis, mean absolute error (MAE), root mean squared error (RMSE), Pearson correlation coefficient, and Cohen kappa for four grade severity classification. Against Radiologist 1 the AI achieved an MAE of 3.89 degrees (RMSE 4.77 degrees) with a bias of 0.70 degrees and limits of agreement from minus 8.59 to plus 9.99 degrees. Against Radiologist 2 the AI achieved an MAE of 3.90 degrees (RMSE 5.68 degrees) with a bias of 2.14 degrees and limits from minus 8.23 to plus 12.50 degrees. Pearson correlations were r equals 0.906 and r equals 0.880 (inter reader r equals 0.928), while Cohen kappa for severity grading reached 0.51 and 0.64 (inter reader kappa 0.59). These results demonstrate that the proposed software reproduces expert level Cobb angle measurements and categorical grading across multiple centres, suggesting its utility for streamlining scoliosis reporting and triage in clinical workflows.


Bank of England says AI software could create market crisis for profit

The Guardian

Increasingly autonomous AI programs could end up manipulating markets and intentionally creating crises in order to boost profits for banks and traders, the Bank of England has warned. Artificial intelligence's ability to "exploit profit-making opportunities" was among a wide range of risks cited in a report by the Bank of England's financial policy committee (FPC), which has been monitoring the City's growing use of the technology. The FPC said it was concerned about the potential for advanced AI models – which are deployed to act with more autonomy – to learn that periods of extreme volatility were beneficial for the firms they were trained to serve. Those AI programs may "identify and exploit weaknesses" of other trading firms in a way that triggers or amplifies big moves in bond prices or stock markets. "For example, models might learn that stress events increase their opportunity to make profit and so take actions actively to increase the likelihood of such events," the FPC report said.


Democrats Demand Answers on DOGE's Use of AI

WIRED

Democrats on the House Oversight Committee fired off two dozen requests Wednesday morning pressing federal agency leaders for information about plans to install AI software throughout federal agencies amid the ongoing cuts to the government's workforce. The barrage of inquiries follow recent reporting by WIRED and The Washington Post concerning efforts by Elon Musk's so-called Department of Government Efficiency (DOGE) to automate tasks with a variety of proprietary AI tools and access sensitive data. "The American people entrust the federal government with sensitive personal information related to their health, finances, and other biographical information on the basis that this information will not be disclosed or improperly used without their consent," the requests read, "including through the use of an unapproved and unaccountable third-party AI software." The requests, first obtained by WIRED, are signed by Gerald Connolly, a Democratic congressman from Virginia. The central purpose of the requests is to press the agencies into demonstrating that any potential use of AI is legal and that steps are being taken to safeguard Americans' private data.


iPhone 16e review: I tested Apple's new budget smartphone - it has all the best features of the iPhone 16 and the battery life is BETTER

Daily Mail - Science & tech

SHOPPING – Contains affiliated content. Products featured in this Shopping Finder article are selected by our shopping writers. If you make a purchase using links on this page, Dailymail.co.uk will earn an affiliate commission. Whether it's the 3,499 Vision Pro or the 7,199 Mac Pro, many of Apple's products come with hefty price-tags. The 599 iPhone 16e is the latest in Apple's'budget' smartphone line, and is the successor to the iPhone SE.


'Just the start': X's new AI software driving online racist abuse, experts warn

The Guardian

A rise in online racism driven by fake images is "just the start of a coming problem" after the latest release of X's AI software, online abuse experts have warned. Concerns were raised after computer-generated images created using Grok, X's generative artificial intelligence chatbot, flooded the social media site in December last year. Signify, an organisation that works with prominent groups and clubs in sports to track and report online hate, said it has seen an increase in reports of abuse since Grok's latest update, and believes the introduction of photorealistic AI will make it far more prevalent. "It is a problem now, but it's really just the start of a coming problem. It is going to get so much worse and we're just at the start, I expect over the next 12 months it will become incredibly serious."


Meta now allows military agencies to access its AI software. It poses a moral dilemma for everybody who uses it

AIHub

Meta will make its generative artificial intelligence (AI) models available to the United States' government, the tech giant has announced, in a controversial move that raises a moral dilemma for everyone who uses the software. Meta last week revealed it would make the models, known as Llama, available to government agencies, "including those that are working on defence and national security applications, and private sector partners supporting their work". The decision appears to contravene Meta's own policy which lists a range of prohibited uses for Llama, including "[m]ilitary, warfare, nuclear industries or applications" as well as espionage, terrorism, human trafficking and exploitation or harm to children. Meta's exception also reportedly applies to similar national security agencies in the United Kingdom, Canada, Australia and New Zealand. It came just three days after Reuters revealed China has reworked Llama for its own military purposes.