Goto

Collaborating Authors

 mercury


'Planetary parade' will see SIX planets align in rare spectacle tonight - here's the best time to spot Mercury, Venus, Jupiter, Saturn, Uranus and Neptune in the night sky

Daily Mail - Science & tech

ROTC students at Old Dominion subdued and killed ISIS-linked gunman who left one dead, two wounded after shouting'Allahu Akbar' and opened fire Horrifying next twist in the Alexander brothers case: MAUREEN CALLAHAN exposes an unthinkable perversion that's been hiding in plain sight Kentucky mother and daughter turn down $26.5MILLION to sell their farms to secretive tech giant that wants to build data center there Hollywood icon who starred in Psycho after Hitchcock dubbed her'my new Grace Kelly' looks incredible at 95 Kylie Jenner's total humiliation in Hollywood: Derogatory rumor leaves her boyfriend's peers'laughing at her' behind her back Tucker Carlson erupts at Trump adviser as she hurls'SLANDER' claim linking him to synagogue shooting Ben Affleck'scores $600m deal' with Netflix to sell his AI film start-up Long hair over 45 is ageing and try-hard. I've finally cut mine off. Alexander brothers' alleged HIGH SCHOOL rape video: Classmates speak out on sickening footage... as creepy unseen photos are exposed Heartbreaking video shows very elderly DoorDash driver shuffle down customer's driveway with coffee order because he is too poor to retire Amber Valletta, 52, was a '90s Vogue model who made movies with Sandra Bullock and Kate Hudson, see her now Model Cindy Crawford, 60, mocked for her'out of touch' morning routine: 'Nothing about this is normal' 'Planetary parade' will see SIX planets align in rare spectacle tonight - here's the best time to spot Mercury, Venus, Jupiter, Saturn, Uranus and Neptune in the night sky Keen astronomers are in for a treat tonight, as a rare'planetary parade' of six planets lights up the night sky. Tonight, Mercury, Venus, Jupiter, Saturn, Uranus, and Neptune will all be visible from Earth. Excitingly, four of these planets will be visible with the naked eye, so you won't need any special equipment to enjoy the spectacle.


Mercury: ACodeEfficiencyBenchmarkforCode LargeLanguageModels

Neural Information Processing Systems

Amidst therecent strides inevaluating LargeLanguage Models forCode (Code LLMs), existing benchmarks havemainly focused onthefunctional correctness of generated code, neglecting the importance of their computational efficiency.


Mercury: A Code Efficiency Benchmark for Code Large Language Models

Neural Information Processing Systems

Amidst the recent strides in evaluating Large Language Models for Code (Code LLMs), existing benchmarks have mainly focused on the functional correctness of generated code, neglecting the importance of their computational efficiency. To fill the gap, we present Mercury, the first code efficiency benchmark for Code LLMs. It comprises 1,889 Python tasks, each accompanied by adequate solutions that serve as real-world efficiency baselines, enabling a comprehensive analysis of the runtime distribution. Based on the distribution, we introduce a new metric Beyond, which computes a runtime-percentile-weighted Pass score to reflect functional correctness and code efficiency simultaneously. On Mercury, leading Code LLMs can achieve 65% on Pass, while less than 50% on Beyond. Given that an ideal Beyond score would be aligned with the Pass score, it indicates that while Code LLMs exhibit impressive capabilities in generating functionally correct code, there remains a notable gap in their efficiency. Finally, our empirical experiments reveal that Direct Preference Optimization (DPO) serves as a robust baseline for enhancing code efficiency compared with Supervised Fine Tuning (SFT), which paves a promising avenue for future exploration of efficient code generation. Our code and data are available on GitHub: https://github.com/Elfsong/Mercury.



Parallel Thinking, Sequential Answering: Bridging NAR and AR for Efficient Reasoning

Ai, Qihang, Jiang, Haiyun

arXiv.org Artificial Intelligence

We study reasoning tasks through a framework that integrates auto-regressive (AR) and non-autoregressive (NAR) language models. AR models, which generate text sequentially, excel at producing coherent outputs but often suffer from slow inference, particularly in reasoning-intensive domains such as mathematics and code, where lengthy chains of thought are required. In contrast, NAR models, such as discrete diffusion models, allow parallel generation and offer substantial speedups, though typically at the cost of reduced output quality. To address these limitations, we introduce a new paradigm in which an NAR model efficiently produces intermediate reasoning traces, which subsequently guide an AR model to deliver precise final answers. Experiments demonstrate that our approach yields significant 26% improvements over strong baselines while substantially reducing inference cost.


Towards Better Correctness and Efficiency in Code Generation

Feng, Yunlong, Xu, Yang, Xu, Xiao, Hui, Binyuan, Lin, Junyang

arXiv.org Artificial Intelligence

While code large language models have demonstrated remarkable progress in code generation, the generated code often exhibits poor runtime efficiency, limiting its practical application in performance-sensitive scenarios. To address this limitation, we propose an efficiency-oriented reinforcement learning framework guided by a novel performance reward. Based on this framework, we take a deeper dive into the code efficiency problem, identifying then proposing methods to overcome key bottlenecks: (1) Dynamic exploration overcomes the static data constraints of offline fine-tuning, enabling the discovery of more efficient code implementations. (2) The error-insensitive reinforcement learning method and high-contrast efficiency signals are crucial for mitigating systematic errors and achieving effective optimization. (3) Online exploration is most effective when starting from a high-correctness baseline, as this allows for efficiency improvements without sacrificing accuracy. With these discoveries, we finally propose a two-stage tuning method, which achieves high and balanced performance across correctness and efficiency. The results of experiments show the effectiveness of the method, which improves code correctness by 10.18\% and runtime efficiency by 7.75\% on a 7B model, achieving performance comparable to much larger model.


Mercury: A Code Efficiency Benchmark for Code Large Language Models

Neural Information Processing Systems

Amidst the recent strides in evaluating Large Language Models for Code (Code LLMs), existing benchmarks have mainly focused on the functional correctness of generated code, neglecting the importance of their computational efficiency. To fill the gap, we present Mercury, the first code efficiency benchmark for Code LLMs. It comprises 1,889 Python tasks, each accompanied by adequate solutions that serve as real-world efficiency baselines, enabling a comprehensive analysis of the runtime distribution. Based on the distribution, we introduce a new metric Beyond, which computes a runtime-percentile-weighted Pass score to reflect functional correctness and code efficiency simultaneously. On Mercury, leading Code LLMs can achieve 65% on Pass, while less than 50% on Beyond.


Hypernym Mercury: Token Optimization Through Semantic Field Constriction And Reconstruction From Hypernyms. A New Text Compression Method

Forrester, Chris, Sulea, Octavia

arXiv.org Artificial Intelligence

Compute optimization using token reduction of LLM prompts is an emerging task in the fields of NLP and next generation, agentic AI. In this white paper, we introduce a novel (patent pending) text representation scheme and a first-of-its-kind word-level semantic compression of paragraphs that can lead to over 90% token reduction, while retaining high semantic similarity to the source text. We explain how this novel compression technique can be lossless and how the detail granularity is controllable. We discuss benchmark results over open source data (i.e. Bram Stoker's Dracula available through Project Gutenberg) and show how our results hold at the paragraph level, across multiple genres and models.


SEANN: A Domain-Informed Neural Network for Epidemiological Insights

Guimbaud, Jean-Baptiste, Plantevit, Marc, Maître, Léa, Cazabet, Rémy

arXiv.org Artificial Intelligence

In epidemiology, traditional statistical methods such as logistic regression, linear regression, and other parametric models are commonly employed to investigate associations between predictors and health outcomes. However, non-parametric machine learning techniques, such as deep neural networks (DNNs), coupled with explainable AI (XAI) tools, offer new opportunities for this task. Despite their potential, these methods face challenges due to the limited availability of high-quality, high-quantity data in this field. To address these challenges, we introduce SEANN, a novel approach for informed DNNs that leverages a prevalent form of domain-specific knowledge: Pooled Effect Sizes (PES). PESs are commonly found in published Meta-Analysis studies, in different forms, and represent a quantitative form of a scientific consensus. By direct integration within the learning procedure using a custom loss, we experimentally demonstrate significant improvements in the generalizability of predictive performances and the scientific plausibility of extracted relationships compared to a domain-knowledge agnostic neural network in a scarce and noisy data setting.


Scientists explain why BepiColombo's mission to Mercury is so tricky

Popular Science

It seems like it should be pretty easy to get to Mercury. The little rocky planet is so much closer to Earth than distant destinations like Jupiter, where we've successfully sent multiple spacecraft. Plus, it doesn't have a crushing atmosphere like our nearest neighbor Venus. But, in fact, it's actually really difficult to reach the innermost planet of our solar system--which makes it that much more impressive that the ESA and JAXA's BepiColombo mission has almost reached Mercury, recently completing its final flyby of the planet before entering orbit next year. Reaching Mercury is such a challenge because "the gravitational pull of the Sun is very strong near Mercury, which makes it difficult for spacecraft to slow down enough to enter orbit around the planet," explains Lina Hadid, staff scientist at CNRS in France and principal investigator of one of BepiColombo's instruments.