alphaevolve
Inside OpenAI's big play for science
An exclusive conversation with Kevin Weil, head of OpenAI for Science, a new in-house team that wants to make scientists more productive. In the three years since ChatGPT's explosive debut, OpenAI's technology has upended a remarkable range of everyday activities at home, at work, in schools--anywhere people have a browser open or a phone out, which is everywhere. Now OpenAI is making an explicit play for scientists. In October, the firm announced that it had launched a whole new team, called OpenAI for Science, dedicated to exploring how its large language models could help scientists and tweaking its tools to support them. The last couple of months have seen a slew of social media posts and academic publications in which mathematicians, physicists, biologists, and others have described how LLMs (and OpenAI's GPT-5 in particular) have helped them make a discovery or nudged them toward a solution they might otherwise have missed. In part, OpenAI for Science was set up to engage with this community.
- North America > United States > Massachusetts (0.04)
- North America > United States > California > Alameda County > Berkeley (0.04)
- Asia > China (0.04)
- Education (0.48)
- Health & Medicine (0.47)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Chatbot (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning > Generative AI (1.00)
What's next for AI in 2026
Our AI writers make their big bets for the coming year--here are five hot trends to watch. In an industry in constant flux, sticking your neck out to predict what's coming next may seem reckless. But for the last few years we've done just that--and we're doing it again. How did we do last time? Here are our big bets for the next 12 months. The last year shaped up as a big one for Chinese open-source models.
- North America > United States > California (0.15)
- Asia > China (0.06)
- North America > United States > Massachusetts (0.04)
- (2 more...)
- Law (1.00)
- Information Technology (1.00)
- Government > Regional Government > North America Government > United States Government (1.00)
- Information Technology > Communications > Social Media (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Chatbot (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (1.00)
ThetaEvolve: Test-time Learning on Open Problems
Wang, Yiping, Su, Shao-Rong, Zeng, Zhiyuan, Xu, Eva, Ren, Liliang, Yang, Xinyu, Huang, Zeyi, He, Xuehai, Ma, Luyao, Peng, Baolin, Cheng, Hao, He, Pengcheng, Chen, Weizhu, Wang, Shuohang, Du, Simon Shaolei, Shen, Yelong
Recent advances in large language models (LLMs) have enabled breakthroughs in mathematical discovery, exemplified by AlphaEvolve, a closed-source system that evolves programs to improve bounds on open problems. However, it relies on ensembles of frontier LLMs to achieve new bounds and is a pure inference system that models cannot internalize the evolving strategies. We introduce ThetaEvolve, an open-source framework that simplifies and extends AlphaEvolve to efficiently scale both in-context learning and Reinforcement Learning (RL) at test time, allowing models to continually learn from their experiences in improving open optimization problems. ThetaEvolve features a single LLM, a large program database for enhanced exploration, batch sampling for higher throughput, lazy penalties to discourage stagnant outputs, and optional reward shaping for stable training signals, etc. ThetaEvolve is the first evolving framework that enable a small open-source model, like DeepSeek-R1-0528-Qwen3-8B, to achieve new best-known bounds on open problems (circle packing and first auto-correlation inequality) mentioned in AlphaEvolve. Besides, across two models and four open tasks, we find that ThetaEvolve with RL at test-time consistently outperforms inference-only baselines, and the model indeed learns evolving capabilities, as the RL-trained checkpoints demonstrate faster progress and better final performance on both trained target task and other unseen tasks. We release our code publicly: https://github.com/ypwang61/ThetaEvolve
- Asia > Middle East > Lebanon > South Governorate > Sidon (0.04)
- North America > United States > Wisconsin > Dane County > Madison (0.04)
- North America > United States > Michigan (0.04)
- North America > United States > California > San Diego County > San Diego (0.04)
GigaEvo: An Open Source Optimization Framework Powered By LLMs And Evolution Algorithms
Khrulkov, Valentin, Galichin, Andrey, Bashkirov, Denis, Vinichenko, Dmitry, Travkin, Oleg, Alferov, Roman, Kuznetsov, Andrey, Oseledets, Ivan
Recent advances in LLM-guided evolutionary computation, particularly AlphaEvolve (Novikov et al., 2025; Georgiev et al., 2025), have demonstrated remarkable success in discovering novel mathematical constructions and solving challenging optimization problems. However, the high-level descriptions in published work leave many implementation details unspecified, hindering reproducibility and further research. In this report we present GigaEvo, an extensible open-source framework that enables researchers to study and experiment with hybrid LLM-evolution approaches inspired by AlphaEvolve. Our system provides modular implementations of key components: MAP-Elites quality-diversity algorithms, asynchronous DAG-based evaluation pipelines, LLM-driven mutation operators with insight generation and bidirectional lineage tracking, and flexible multi-island evolutionary strategies. In order to assess reproducibility and validate our implementation we evaluate GigaEvo on challenging problems from the AlphaEvolve paper: Heil-bronn triangle placement, circle packing in squares, and high-dimensional kissing numbers. The framework emphasizes modularity, concurrency, and ease of experimentation, enabling rapid prototyping through declarative configuration. We provide detailed descriptions of system architecture, implementation decisions, and experimental methodology to support further research in LLM-driven evolutionary methods. The recent paper (Novikov et al., 2025) introduced AlphaEvolve, a framework that combines large language model (LLM) code generation with evolutionary computation, achieving state-of-the-art results on challenging algorithmic and mathematical problems.
- Europe > Russia > Central Federal District > Moscow Oblast > Moscow (0.04)
- Asia > Russia (0.04)
- Asia > Middle East > Jordan (0.04)
Mathematicians say Google's AI tools are supercharging their research
Mathematicians say Google's AI tools are supercharging their research AI tools developed by Google DeepMind are surprisingly effective at assisting mathematical research and could usher in a wave of AI-powered mathematical discovery at a previously unseen scale, say mathematicians who have tested the technology. In May, Google announced an AI system called AlphaEvolve that could find new algorithms and mathematical formulae. The system works by exploring many possible solutions, produced by Google's AI chatbot Gemini. Crucially, though, these are fed to a separate AI evaluator that can filter out the nonsensical solutions that a chatbot inevitably generates . At the time, Google researchers tested AlphaEvolve on more than 50 open mathematical problems and found that, in three-quarters of cases, the system could rediscover the best-known solutions found by humans.
- North America > United States > California > Los Angeles County > Los Angeles (0.15)
- North America > United States > Rhode Island (0.05)
- North America > United States > Pennsylvania (0.05)
- (3 more...)
Human outsmarts Google DeepMind AI, solving centuries-old 'kissing problem'
How many spheres can be arranged so that every one'kisses' a single rounded shape in the center? Breakthroughs, discoveries, and DIY tips sent every weekday. A human has outkissed one of Google's superpowered artificial intelligence systems . Instead, this win is in the intellectual realm of advanced mathematics . While largely conceptual in nature, the ramifications could soon help boost advancements in telecommunications and satellite arrays.
Scientific Algorithm Discovery by Augmenting AlphaEvolve with Deep Research
Liu, Gang, Zhu, Yihan, Chen, Jie, Jiang, Meng
Large language models hold promise as scientific assistants, yet existing agents either rely solely on algorithm evolution or on deep research in isolation, both of which face critical limitations. Pure algorithm evolution, as in AlphaEvolve, depends only on the internal knowledge of LLMs and quickly plateaus in complex domains, while pure deep research proposes ideas without validation, resulting in unrealistic or unimplementable solutions. We present DeepEvolve, an agent that integrates deep research with algorithm evolution, uniting external knowledge retrieval, cross-file code editing, and systematic debugging under a feedback-driven iterative loop. Each iteration not only proposes new hypotheses but also refines, implements, and tests them, avoiding both shallow improvements and unproductive over-refinements. Across nine benchmarks in chemistry, mathematics, biology, materials, and patents, DeepEvolve consistently improves the initial algorithm, producing executable new algorithms with sustained gains. By bridging the gap between unguided evolution and research without grounding, DeepEvolve provides a reliable framework for advancing scientific algorithm discovery. Our code is available at https://github.com/liugangcode/deepevolve.
- North America > United States > Michigan (0.04)
- Asia > Middle East > Jordan (0.04)
The Mathematician's Assistant: Integrating AI into Research Practice
The rapid development of artificial intelligence (AI), marked by breakthroughs like 'AlphaEvolve' and 'Gemini Deep Think', is beginning to offer powerful new tools that have the potential to significantly alter the research practice in many areas of mathematics. This paper explores the current landscape of publicly accessible large language models (LLMs) in a mathematical research context, based on developments up to August 2, 2025. Our analysis of recent benchmarks, such as MathArena and the Open Proof Corpus (Balunović et al., 2025; Dekoninck et al., 2025), reveals a complex duality: while state-of-the-art models demonstrate strong abilities in solving problems and evaluating proofs, they also exhibit systematic flaws, including a lack of self-critique and a model depending discrepancy between final-answer accuracy and full-proof validity. Based on these findings, we propose a durable framework for integrating AI into the research workflow, centered on the principle of the augmented mathematician. In this model, the AI functions as a copilot under the critical guidance of the human researcher, an approach distilled into five guiding principles for effective and responsible use. We then systematically explore seven fundamental ways AI can be applied across the research lifecycle, from creativity and ideation to the final writing process, demonstrating how these principles translate into concrete practice. We conclude that the primary role of AI is currently augmentation rather than automation. This requires a new skill set focused on strategic prompting, critical verification, and methodological rigor in order to effectively use these powerful tools.
- Europe > Switzerland > Zürich > Zürich (0.04)
- North America > United States > Virginia (0.04)
- North America > United States > California > Santa Clara County > Stanford (0.04)
- Europe > Germany (0.04)
- Research Report > New Finding (0.67)
- Personal > Honors (0.67)
- Research Report > Promising Solution (0.66)
- Information Technology > Security & Privacy (1.00)
- Education > Educational Setting > Higher Education (0.46)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Chatbot (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (1.00)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Agents (0.93)
AlphaEvolve: A coding agent for scientific and algorithmic discovery
Novikov, Alexander, Vũ, Ngân, Eisenberger, Marvin, Dupont, Emilien, Huang, Po-Sen, Wagner, Adam Zsolt, Shirobokov, Sergey, Kozlovskii, Borislav, Ruiz, Francisco J. R., Mehrabian, Abbas, Kumar, M. Pawan, See, Abigail, Chaudhuri, Swarat, Holland, George, Davies, Alex, Nowozin, Sebastian, Kohli, Pushmeet, Balog, Matej
In this white paper, we present AlphaEvolve, an evolutionary coding agent that substantially enhances capabilities of state-of-the-art LLMs on highly challenging tasks such as tackling open scientific problems or optimizing critical pieces of computational infrastructure. AlphaEvolve orchestrates an autonomous pipeline of LLMs, whose task is to improve an algorithm by making direct changes to the code. Using an evolutionary approach, continuously receiving feedback from one or more evaluators, AlphaEvolve iteratively improves the algorithm, potentially leading to new scientific and practical discoveries. We demonstrate the broad applicability of this approach by applying it to a number of important computational problems. When applied to optimizing critical components of large-scale computational stacks at Google, AlphaEvolve developed a more efficient scheduling algorithm for data centers, found a functionally equivalent simplification in the circuit design of hardware accelerators, and accelerated the training of the LLM underpinning AlphaEvolve itself. Furthermore, AlphaEvolve discovered novel, provably correct algorithms that surpass state-of-the-art solutions on a spectrum of problems in mathematics and computer science, significantly expanding the scope of prior automated discovery methods (Romera-Paredes et al., 2023). Notably, AlphaEvolve developed a search algorithm that found a procedure to multiply two $4 \times 4$ complex-valued matrices using $48$ scalar multiplications; offering the first improvement, after 56 years, over Strassen's algorithm in this setting. We believe AlphaEvolve and coding agents like it can have a significant impact in improving solutions of problems across many areas of science and computation.
- Asia > Middle East > Lebanon > South Governorate > Sidon (0.04)
- North America > United States > Texas > Harris County > Houston (0.04)
- North America > United States > New York > New York County > New York City (0.04)
- (6 more...)
- Research Report > New Finding (0.67)
- Research Report > Promising Solution (0.66)
- Health & Medicine > Pharmaceuticals & Biotechnology (0.67)
- Information Technology > Services (0.54)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Search (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Evolutionary Systems (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.93)
Google DeepMind's AI Agent Dreams Up Algorithms Beyond Human Expertise
A key question in artificial intelligence is how often models go beyond just regurgitating and remixing what they have learned and produce truly novel ideas or insights. A new project from Google DeepMind shows that with a few clever tweaks these models can at least surpass human expertise designing certain types of algorithms--including ones that are useful for advancing AI itself. The company's latest AI project, called AlphaEvolve, combines the coding skills of its Gemini AI model with a method for testing the effectiveness of new algorithms and an evolutionary method for producing new designs. AlphaEvolve came up with more efficient algorithms for several kinds of computation, including a method for calculations involving matrices that betters an approach called the Strassen algorithm that has been relied upon for 56 years. The new approach improves the computational efficiency by reducing the number of calculations required to produce a result.