Goto

Collaborating Authors

 butterfly


9 new butterflies discovered in old museum archives

Popular Science

The team even extracted DNA from a tiny 100-year-old butterfly leg. Breakthroughs, discoveries, and DIY tips sent every weekday. When you think of butterflies, chances are you imagine unmistakable insects with bright, bold wings. But it turns out that individual butterfly species are sometimes shockingly difficult to tell apart. "Thanks to the genetic revolution and the collaboration of researchers and museums in various countries led by London's Natural History Museum, century-old butterflies are now speaking to us," Christophe Faynel, an entomologist at the Société entomologique Antilles Guyane, said in a statement .


Where did I put it? Loss of vital crypto key voids election

New Scientist

Feedback is entertained by the commotion at the International Association for Cryptologic Research's recent elections, where results could not be decrypted after an honest but unfortunate human mistake The phrase "you couldn't make it up", Feedback feels, is often misunderstood. It doesn't mean there are limits to the imagination, but rather that there are some developments you can't include in a fictional story because people would say "oh come on, that would never happen". The trouble is, those people are wrong, because real life is frequently ridiculous. In the world of codes and ciphers, one of the more important organisations is the International Association for Cryptologic Research, described as " a non-profit organization devoted to supporting the promotion of the science of cryptology ". The IACR recently held elections to choose new officers and directors and to tweak its bylaws.


BotaCLIP: Contrastive Learning for Botany-Aware Representation of Earth Observation Data

Cerna, Selene, Si-Moussi, Sara, Thuiller, Wilfried, Hendrikx, Hadrien, Miele, Vincent

arXiv.org Artificial Intelligence

Foundation models have demonstrated a remarkable ability to learn rich, transferable representations across diverse modalities such as images, text, and audio. In modern machine learning pipelines, these representations often replace raw data as the primary input for downstream tasks. In this paper, we address the challenge of adapting a pre-trained foundation model to inject domain-specific knowledge, without retraining from scratch or incurring significant computational costs. To this end, we introduce BotaCLIP, a lightweight multimodal contrastive framework that adapts a pre-trained Earth Observation foundation model (DOFA) by aligning high-resolution aerial imagery with botanical relevés. Unlike generic embeddings, BotaCLIP internalizes ecological structure through contrastive learning with a regularization strategy that mitigates catastrophic forgetting. Once trained, the resulting embeddings serve as transferable representations for downstream predictors. Motivated by real-world applications in biodiversity modeling, we evaluated BotaCLIP representations in three ecological tasks: plant presence prediction, butterfly occurrence modeling, and soil trophic group abundance estimation. The results showed consistent improvements over those derived from DOFA and supervised baselines. More broadly, this work illustrates how domain-aware adaptation of foundation models can inject expert knowledge into data-scarce settings, enabling frugal representation learning.



The '10 Martini' Proof Connects Quantum Mechanics With Infinitely Intricate Mathematical Structures

WIRED

The proof, known to be so hard that a mathematician once offered 10 martinis to whoever could figure it out, uses number theory to explain quantum fractals. In 1974, five years before he wrote his Pulitzer Prize-winning book, Douglas Hofstadter was a graduate student in physics at the University of Oregon. When his doctoral adviser went on sabbatical to Regensburg, Germany, Hofstadter tagged along, hoping to practice his German. The pair joined a group of brilliant theoretical physicists who were agonizing over a particular problem in quantum theory. They wanted to determine the energy levels of an electron in a crystal grid placed near a magnet. Hofstadter was the odd one out, unable to follow the others' line of thought. "Part of my luck was that I couldn't keep up with them," he said.



SONAR-LLM: Autoregressive Transformer that Thinks in Sentence Embeddings and Speaks in Tokens

Dragunov, Nikita, Rahmatullaev, Temurbek, Goncharova, Elizaveta, Kuznetsov, Andrey, Razzhigaev, Anton

arXiv.org Artificial Intelligence

The recently proposed Large Concept Model (LCM) generates text by predicting a sequence of sentence-level embeddings and training with either mean-squared error or diffusion objectives. We present SONAR-LLM, a decoder-only transformer that "thinks" in the same continuous SONAR embedding space, yet is supervised through token-level cross-entropy propagated via the frozen SONAR decoder. This hybrid objective retains the semantic abstraction of LCM while eliminating its diffusion sampler and restoring a likelihood-based training signal. Across model sizes from 39M to 1.3B parameters, SONAR-LLM attains competitive generation quality. We report scaling trends, ablations, benchmark results, and release the complete training code and all pretrained checkpoints to foster reproducibility and future research.


Computational Advantages of Multi-Grade Deep Learning: Convergence Analysis and Performance Insights

Fang, Ronglong, Xu, Yuesheng

arXiv.org Artificial Intelligence

Multi-grade deep learning (MGDL) has been shown to significantly outperform the standard single-grade deep learning (SGDL) across various applications. This work aims to investigate the computational advantages of MGDL focusing on its performance in image regression, denoising, and deblurring tasks, and comparing it to SGDL. We establish convergence results for the gradient descent (GD) method applied to these models and provide mathematical insights into MGDL's improved performance. In particular, we demonstrate that MGDL is more robust to the choice of learning rate under GD than SGDL. Furthermore, we analyze the eigenvalue distributions of the Jacobian matrices associated with the iterative schemes arising from the GD iterations, offering an explanation for MGDL's enhanced training stability.


Distributed Butterfly Analysis using Mobile Agents

Chand, Prabhat Kumar, Das, Apurba, Molla, Anisur Rahaman

arXiv.org Artificial Intelligence

Butterflies, or 4-cycles in bipartite graphs, are crucial for identifying cohesive structures and dense subgraphs. While agent-based data mining is gaining prominence, its application to bipartite networks remains relatively unexplored. We propose distributed, agent-based algorithms for \emph{Butterfly Counting} in a bipartite graph $G((A,B),E)$. Agents first determine their respective partitions and collaboratively construct a spanning tree, electing a leader within $O(n \log λ)$ rounds using only $O(\log λ)$ bits per agent. A novel meeting mechanism between adjacent agents improves efficiency and eliminates the need for prior knowledge of the graph, requiring only the highest agent ID $λ$ among the $n$ agents. Notably, our techniques naturally extend to general graphs, where leader election and spanning tree construction maintain the same round and memory complexities. Building on these foundations, agents count butterflies per node in $O(Δ)$ rounds and compute the total butterfly count of $G$ in $O(Δ+\min\{|A|,|B|\})$ rounds.


How Do People Revise Inconsistent Beliefs? Examining Belief Revision in Humans with User Studies

Vasileiou, Stylianos Loukas, Rago, Antonio, Martinez, Maria Vanina, Yeoh, William

arXiv.org Artificial Intelligence

Understanding how humans revise their beliefs in light of new information is crucial for developing AI systems which can effectively model, and thus align with, human reasoning. While theoretical belief revision frameworks rely on a set of principles that establish how these operations are performed, empirical evidence from cognitive psychology suggests that people may follow different patterns when presented with conflicting information. In this paper, we present three comprehensive user studies showing that people consistently prefer explanation-based revisions, i.e., those which are guided by explanations, that result in changes to their belief systems that are not necessarily captured by classical belief change theory. Our experiments systematically investigate how people revise their beliefs with explanations for inconsistencies, whether they are provided with them or left to formulate them themselves, demonstrating a robust preference for what may seem non-minimal revisions across different types of scenarios. These findings have implications for AI systems designed to model human reasoning or interact with humans, suggesting that such systems should accommodate explanation-based, potentially non-minimal belief revision operators to better align with human cognitive processes.