trustworthy
AI Leaders Discuss How to Foster Responsible Innovation at TIME100 Roundtable in Davos
Javed is a senior editor at TIME, based in the London bureau. Javed is a senior editor at TIME, based in the London bureau. Leaders from across the tech sector, academia, and beyond gathered to explore how to implement responsible AI and ensure safeguarding while fostering innovation, at a roundtable convened by TIME in Davos, Switzerland, on Jan 21. In a wide-ranging conversation, participants in the roundtable, hosted by TIME CEO Jess Sibley, discussed topics including the impact of AI on children's development and safety, how to regulate the technology, and how to better train models to ensure they don't harm humans. Discussing the safety of children, Jonathan Haidt, professor of ethical leadership at NYU Stern and author of said that parents shouldn't focus on restricting their child's exposure entirely but on the habits they form.
- Europe > Switzerland (0.25)
- North America > United States (0.15)
- Europe > France (0.07)
- (5 more...)
- Information Technology > Services (0.32)
- Government > Military (0.30)
The overlooked driver of digital transformation
Clear, reliable audio is no longer optional, say Genevieve Juillard, CEO of IDC, and Chris Schyvinck, president and CEO at Shure. When business leaders talk about digital transformation, their focus often jumps straight to cloud platforms, AI tools, or collaboration software. Yet, one of the most fundamental enablers of how organizations now work, and how employees experience that work, is often overlooked: audio. As Genevieve Juillard, CEO of IDC, notes, the shift to hybrid collaboration made every space, from corporate boardrooms to kitchen tables, meeting-ready almost overnight. In the scramble, audio quality often lagged, creating what research now shows is more than a nuisance. Poor sound can alter how speakers are perceived, making them seem less credible or even less trustworthy. Audio is the gatekeeper of meaning," stresses Julliard. "If people can't hear clearly, they can't understand you. And if they can't understand you, they can't trust you, and they can't act on what you said. And no amount of sharp video can fix that. For Shure, which has spent a century advancing sound technology, the implications extend far beyond convenience.
- North America > United States > Massachusetts (0.04)
- Europe > United Kingdom > England > East Sussex > Brighton (0.04)
- Asia > China > Beijing > Beijing (0.04)
- Health & Medicine (0.68)
- Information Technology (0.48)
Generative AI hype distracts us from AI's more important breakthroughs
It's a seductive distraction from the advances in AI that are most likely to improve or even save your life On April 28, 2022, at a highly anticipated concert in Spokane, Washington, the musician Paul McCartney astonished his audience with a groundbreaking application of AI: He began to perform with a lifelike depiction of his long-deceased musical partner, John Lennon. Using recent advances in audio and video processing, engineers had taken the pair's final performance (London, 1969), separated Lennon's voice and image from the original mix and restored them with lifelike clarity. For years, researchers like me had taught machines to "see" and "hear" in order to make such a moment possible. As McCartney and Lennon appeared to reunite across time and space, the arena fell silent; many in the crowd began to cry. As an AI scientist and lifelong Beatles fan, I felt profound gratitude that we could experience this truly life-changing moment. Later that year, the world was captivated by another major breakthrough: AI conversation.
- North America > United States > Washington > Spokane County > Spokane (0.25)
- North America > United States > Massachusetts (0.05)
OpenAI has trained its LLM to confess to bad behavior
Large language models often lie and cheat. We can't stop that--but we can make them own up. OpenAI is testing another new way to expose the complicated processes at work inside large language models. Researchers at the company can make an LLM produce what they call a confession, in which the model explains how it carried out a task and (most of the time) owns up to any bad behavior. Figuring out why large language models do what they do--and in particular why they sometimes appear to lie, cheat, and deceive--is one of the hottest topics in AI right now. If this multitrillion-dollar technology is to be deployed as widely as its makers hope it will be, it must be made more trustworthy.
- North America > United States > Massachusetts (0.05)
- Asia > China (0.05)
Dynamic Logic of Trust-Based Beliefs
Jiang, Junli, Naumov, Pavel, Zhang, Wenxuan
Traditionally, an agent's beliefs would come from what the agent can see, hear, or sense. In the modern world, beliefs are often based on the data available to the agents. In this work, we investigate a dynamic logic of such beliefs that incorporates public announcements of data. The main technical contribution is a sound and complete axiomatisation of the interplay between data-informed beliefs and data announcement modalities. We also describe a non-trivial polynomial model checking algorithm for this logical system.
- North America > United States (0.47)
- Europe > United Kingdom > England > Merseyside > Liverpool (0.04)
- Europe > United Kingdom > England > Hampshire > Southampton (0.04)
- (2 more...)
Game of Trust: How Trustworthy Does Your Blockchain Think You Are?
Drineas, Petros, Nema, Rohit, Ostrovsky, Rafail, Zikas, Vassilis
We investigate how a blockchain can distill the collective belief of its nodes regarding the trustworthiness of a (sub)set of nodes into a {\em reputation system} that reflects the probability of correctly performing a task. To address this question, we introduce a framework that breaks it down into two sub-problems: 1. (Information Extraction): How can the system distill trust information from a function of the nodes' true beliefs? 2. (Incentive Design): How can we incentivize nodes to truthfully report such information? To tackle the first sub-problem, we adapt, in a non-trivial manner, the well-known PageRank algorithm to our problem. For the second, we define a new class of games, called Trustworthy Reputation games (TRep games), which aim to extract the collective beliefs on trust from the actions of rational participants. We then propose a concrete TRep game whose utility function leverages Personalized PageRank and can be instantiated through a straightforward blockchain rewards mechanism. Building on this, we show how the TRep game enables the design of a reputation system. Such systems can enhance the robustness, scalability, and efficiency of blockchain and DeFi solutions. For instance, we demonstrate how such a system can be used within a Proof-of-Reputation blockchain.
- North America > United States > New York > New York County > New York City (0.14)
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.04)
- North America > United States > Massachusetts > Middlesex County > Cambridge (0.04)
- (8 more...)
- Information Technology > Security & Privacy (1.00)
- Banking & Finance > Trading (1.00)
- Leisure & Entertainment > Games (0.68)
AI now sounds more like us – should we be concerned?
AI now sounds more like us - should we be concerned? Several wealthy Italian businessmen received a surprising phone call earlier this year. The speaker, who sounded just like Defence Minister Guido Crosetto, had a special request: Please send money to help us free kidnapped Italian journalists in the Middle East. But it was not Crosetto at the end of the line. He only learned about the calls when several of the targeted businessmen contacted him about them.
- Europe > Middle East (0.25)
- Africa > Middle East (0.25)
- South America (0.05)
- (10 more...)
- Government (1.00)
- Media > News (0.70)
- Law Enforcement & Public Safety > Crime Prevention & Enforcement (0.48)
The SMeL Test: A simple benchmark for media literacy in language models
Ahdritz, Gustaf, Kleiman, Anat
The internet is rife with unattributed, deliberately misleading, or otherwise untrustworthy content. Though large language models (LLMs) are often tasked with autonomous web browsing, the extent to which they have learned the simple heuristics human researchers use to navigate this noisy environment is not currently known. In this paper, we introduce the Synthetic Media Literacy Test (SMeL Test), a minimal benchmark that tests the ability of language models to actively filter out untrustworthy information in context. We benchmark a variety of commonly used instruction-tuned LLMs, including reasoning models, and find that no model consistently succeeds; while reasoning in particular is associated with higher scores, even the best API model we test hallucinates up to 70% of the time. Remarkably, larger and more capable models do not necessarily outperform their smaller counterparts. We hope our work sheds more light on this important form of hallucination and guides the development of new methods to combat it.
- North America > Canada (0.04)
- Asia > Singapore (0.04)
- Asia > Indonesia > Bali (0.04)
- (9 more...)
- Government (0.93)
- Media > News (0.68)
Trustworthiness Preservation by Copies of Machine Learning Systems
Ceragioli, Leonardo, Primiero, Giuseppe
A common practice of ML systems development concerns the training of the same model under different data sets, and the use of the same (training and test) sets for different learning models. The first case is a desirable practice for identifying high quality and unbiased training conditions. The latter case coincides with the search for optimal models under a common dataset for training. These differently obtained systems have been considered akin to copies. In the quest for responsible AI, a legitimate but hardly investigated question is how to verify that trustworthiness is preserved by copies. In this paper we introduce a calculus to model and verify probabilistic complex queries over data and define four distinct notions: Justifiably, Equally, Weakly and Almost Trustworthy which can be checked analysing the (partial) behaviour of the copy with respect to its original. We provide a study of the relations between these notions of trustworthiness, and how they compose with each other and under logical operations. The aim is to offer a computational tool to check the trustworthiness of possibly complex systems copied from an original whose behavour is known.
- Europe > Italy > Lazio > Rome (0.04)
- Europe > United Kingdom > England > Oxfordshire > Oxford (0.04)
- Europe > Estonia > Harju County > Tallinn (0.04)
- Asia > Japan (0.04)
- Health & Medicine (0.73)
- Banking & Finance (0.46)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Logic & Formal Reasoning (1.00)
- Information Technology > Artificial Intelligence > Machine Learning (1.00)
- Information Technology > Artificial Intelligence > Issues > Social & Ethical Issues (1.00)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Uncertainty (0.93)
The Most-Cited Computer Scientist Has a Plan to Make AI More Trustworthy
On June 3, Yoshua Bengio, the world's most-cited computer scientist, announced the launch of LawZero, a nonprofit that aims to create "safe by design" AI by pursuing a fundamentally different approach to major tech companies. Players like OpenAI and Google are investing heavily in AI agents--systems that not only answer queries and generate images, but can craft plans and take actions in the world. The goal of these companies is to create virtual employees that can do practically any job a human can, known in the tech industry as artificial general intelligence, or AGI. Executives like Google DeepMind's CEO Demis Hassabis point to AGI's potential to solve climate change or cure disease as a motivator for its development. Bengio, however, says we don't need agentic systems to reap AI's rewards--it's a false choice.