superhuman ai
The Doomers Who Insist AI Will Kill Us All
The subtitle of the doom bible to be published by AI extinction prophets Eliezer Yudkowsky and Nate Soares later this month is "Why superhuman AI would kill us all." But it really should be "Why superhuman AI WILL kill us all," because even the coauthors don't believe that the world will take the necessary measures to stop AI from eliminating all non-super humans. The book is beyond dark, reading like notes scrawled in a dimly lit prison cell the night before a dawn execution. When I meet these self-appointed Cassandras, I ask them outright if they believe that they personally will meet their ends through some machination of superintelligence. The answers come promptly: "yeah" and "yup."
Scalable Oversight for Superhuman AI via Recursive Self-Critiquing
Wen, Xueru, Lou, Jie, Lu, Xinyu, Yang, Junjie, Liu, Yanjiang, Lu, Yaojie, Zhang, Debing, XingYu, null
As AI capabilities increasingly surpass human proficiency in complex tasks, current alignment techniques including SFT and RLHF face fundamental challenges in ensuring reliable oversight. These methods rely on direct human assessment and become untenable when AI outputs exceed human cognitive thresholds. In response to this challenge, we explore two hypotheses: (1) critique of critique can be easier than critique itself, extending the widely-accepted observation that verification is easier than generation to the critique domain, as critique itself is a specialized form of generation; (2) this difficulty relationship is recursively held, suggesting that when direct evaluation is infeasible, performing high-order critiques (e.g., critique of critique of critique) offers a more tractable supervision pathway. To examine these hypotheses, we perform Human-Human, Human-AI, and AI-AI experiments across multiple tasks. Our results demonstrate encouraging evidence supporting these hypotheses and suggest that recursive self-critiquing is a promising direction for scalable oversight.
- North America > United States > Minnesota > Hennepin County > Minneapolis (0.14)
- Antarctica (0.04)
- Africa (0.04)
- (3 more...)
How to Hit Pause on AI Before It's Too Late
Only 16 months have passed, but the release of ChatGPT back in November 2022 feels already like ancient AI history. Hundreds of billions of dollars, both public and private, are pouring into AI. Thousands of AI-powered products have been created, including the new GPT-4o just this week. Everyone from students to scientists now use these large language models. Our world, and in particular the world of AI, has decidedly changed.
Elon Musk predicts superhuman AI will be smarter than people next year
Superhuman artificial intelligence that is smarter than anyone on Earth could exist next year, Elon Musk has said, unless the sector's power and computing demands become unsustainable before then. The prediction is a sharp tightening of an earlier claim from the multibillionaire, that superintelligent AI would exist by 2029. Whereas "superhuman" is generally defined as being smarter than any individual human at any specific task, superintelligent is often defined instead as being smarter than every human's combined ability at any task. "My guess is that we'll have AI that is smarter than any one human probably around the end of next year," Musk said in a live streamed interview on his social network X. That prediction was made with the caveat that increasing demands for power and shortages of the most powerful AI training chips could limit their capability in the near term.
- North America > United States > New York (0.06)
- North America > United States > California > Los Angeles County > Los Angeles (0.06)
- Information Technology (0.37)
- Law (0.33)
Exploring the Constraints on Artificial General Intelligence: A Game-Theoretic No-Go Theorem
The emergence of increasingly sophisticated artificial intelligence (AI) systems have sparked intense debate among researchers, policymakers, and the public due to their potential to surpass human intelligence and capabilities in all domains. In this paper, I propose a game-theoretic framework that captures the strategic interactions between a human agent and a potential superhuman machine agent. I identify four key assumptions: Strategic Unpredictability, Access to Machine's Strategy, Rationality, and Superhuman Machine. The main result of this paper is an impossibility theorem: these four assumptions are inconsistent when taken together, but relaxing any one of them results in a consistent set of assumptions. Two straightforward policy recommendations follow: first, policymakers should control access to specific human data to maintain Strategic Unpredictability; and second, they should grant select AI researchers access to superhuman machine research to ensure Access to Machine's Strategy holds. My analysis contributes to a better understanding of the context that can shape the theoretical development of superhuman AI.
- Europe > United Kingdom > England > Oxfordshire > Oxford (0.04)
- North America > United States > Texas (0.04)
- North America > United States > New York (0.04)
- (3 more...)
- Leisure & Entertainment > Games > Chess (0.95)
- Government (0.86)
- Information Technology (0.68)
- Information Technology > Artificial Intelligence > Issues > Social & Ethical Issues (0.93)
- Information Technology > Artificial Intelligence > Cognitive Science (0.70)
- Information Technology > Artificial Intelligence > Machine Learning > Reinforcement Learning (0.68)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks (0.68)
Superhuman Artificial Intelligence Can Improve Human Decision Making by Increasing Novelty
Shin, Minkyu, Kim, Jin, van Opheusden, Bas, Griffiths, Thomas L.
How will superhuman artificial intelligence (AI) affect human decision making? And what will be the mechanisms behind this effect? We address these questions in a domain where AI already exceeds human performance, analyzing more than 5.8 million move decisions made by professional Go players over the past 71 years (1950-2021). To address the first question, we use a superhuman AI program to estimate the quality of human decisions across time, generating 58 billion counterfactual game patterns and comparing the win rates of actual human decisions with those of counterfactual AI decisions. We find that humans began to make significantly better decisions following the advent of superhuman AI. We then examine human players' strategies across time and find that novel decisions (i.e., previously unobserved moves) occurred more frequently and became associated with higher decision quality after the advent of superhuman AI. Our findings suggest that the development of superhuman AI programs may have prompted human players to break away from traditional strategies and induced them to explore novel moves, which in turn may have improved their decision-making.
- Asia > China > Hong Kong (0.04)
- North America > United States > New Jersey > Mercer County > Princeton (0.04)
- North America > United States > Illinois > Cook County > Chicago (0.04)
- (2 more...)
Humans have improved at Go since AIs became best in the world
AIs can beat the world's best players at the board game Go but humans are starting to improve too. An analysis of millions of Go moves has found that professional players have been making better and more original game choices since Go-playing AIs overtook humans. Before 2016, AIs couldn't beat the world's best Go players. But this changed with an AI called AlphaGo developed by London-based research firm DeepMind. AlphaGo defeated multiple Go champions, including the then number one ranked human player.
- North America > United States > California > Alameda County > Berkeley (0.06)
- Asia > China > Hong Kong (0.06)
Benefits & Risks of Artificial Intelligence - Future of Life Institute
"Everything we love about civilization is a product of intelligence, so amplifying our human intelligence with artificial intelligence has the potential of helping civilization flourish like never before - as long as we manage to keep the technology beneficial." From SIRI to self-driving cars, artificial intelligence (AI) is progressing rapidly. While science fiction often portrays AI as robots with human-like characteristics, AI can encompass anything from Google's search algorithms to IBM's Watson to autonomous weapons. Artificial intelligence today is properly known as narrow AI (or weak AI), in that it is designed to perform a narrow task (e.g. However, the long-term goal of many researchers is to create general AI (AGI or strong AI).
- Information Technology > Robotics & Automation (0.49)
- Energy > Power Industry (0.47)
- Transportation > Passenger (0.35)
- Transportation > Ground > Road (0.35)
The Myth of a Superhuman AI
I've heard that in the future computerized AIs will become so much smarter than us that they will take all our jobs and resources, and humans will go extinct. That's the most common question I get whenever I give a talk about AI. The questioners are earnest; their worry stems in part from some experts who are asking themselves the same thing. These folks are some of the smartest people alive today, such as Stephen Hawking, Elon Musk, Max Tegmark, Sam Harris, and Bill Gates, and they believe this scenario very likely could be true. Recently at a conference convened to discuss these AI issues, a panel of nine of the most informed gurus on AI all agreed this superhuman intelligence was inevitable and not far away.
Artificial general intelligence: Are we close, and does it even make sense to try?
The idea of artificial general intelligence as we know it today starts with a dot-com blowout on Broadway. Twenty years ago--before Shane Legg clicked with neuroscience postgrad Demis Hassabis over a shared fascination with intelligence; before the pair hooked up with Hassabis's childhood friend Mustafa Suleyman, a progressive activist, to spin that fascination into a company called DeepMind; before Google bought that company for more than half a billion dollars four years later--Legg worked at a startup in New York called Webmind, set up by AI researcher Ben Goertzel. Today the two men represent two very different branches of the future of artificial intelligence, but their roots reach back to common ground. Even for the heady days of the dot-com bubble, Webmind's goals were ambitious. Goertzel wanted to create a digital baby brain and release it onto the internet, where he believed it would grow up to become fully self-aware and far smarter than humans.
- Banking & Finance (0.35)
- Health & Medicine (0.35)
- Leisure & Entertainment > Games (0.31)