Goto

Collaborating Authors

 owl


Ancient bees laid eggs inside bones

Popular Science

A 20,000 year old fossil uncovered in a tarantula-filled cave has paleontologists stunned. Breakthroughs, discoveries, and DIY tips sent every weekday. Bees are frequently associated with large queen-serving colonies featuring hundreds if not thousands of insects . They lay their eggs in small cavities, and they leave pollen for the larvae to eat," explained paleontologist Lazasro Viñola López . "Some bee species burrow holes in wood or in the ground, or use empty structures for nests." Viñola López, a researcher at Chicago's Field Museum, added that some European and African species even construct nests inside vacant snail shells. That said, a beehive inside a bone is a new one even for seasoned researchers. Estimated to be around 20,000 years old, this newly discovered specimen is also the first known example of such a home, past or present. The findings are detailed in a study published on December 16 in the journal . Researchers located the unique find while exploring the many limestone caves that dot the southern Dominican Republic. Sinkholes are common across the Caribbean island of Hispaniola, and are often so well sheltered from the elements that they function like underground time capsules. These windows into the past are largely thanks to the work of the island's owls . The predatory birds often make their nests inside these caves, where they regularly cough up owl pellets filled with the undigested bones of their prey. Over thousands of years, these layers of bones fossilize atop one another across carbonate layers created from rainy periods. "The initial descent into the cave isn't too deep-we would tie a rope to the side and then rappel down," Viñola López said. "If you go in at night, you see the eyes of the tarantulas that live inside." After proceeding past the large spiders through about 33 feet of underground tunnel, the paleontologists began finding various fossils. Many belonged to rodents, but there were also bones from birds, reptiles, and even sloths for a total of over 50 different animal species. "We think that this was a cave where owls lived for many generations, maybe for hundreds or thousands of years," said Viñola López. "The owls would go out and hunt, and then come back to the cave and throw up pellets.


CMOMgen: Complex Multi-Ontology Alignment via Pattern-Guided In-Context Learning

Silva, Marta Contreiras, Faria, Daniel, Pesquita, Catia

arXiv.org Artificial Intelligence

Constructing comprehensive knowledge graphs requires the use of multiple ontologies in order to fully contextualize data into a domain. Ontology matching finds equivalences between concepts interconnecting ontologies and creating a cohesive semantic layer. While the simple pairwise state of the art is well established, simple equivalence mappings cannot provide full semantic integration of related but disjoint ontologies. Complex multi-ontology matching (CMOM) aligns one source entity to composite logical expressions of multiple target entities, establishing more nuanced equivalences and provenance along the ontological hierarchy. We present CMOMgen, the first end-to-end CMOM strategy that generates complete and semantically sound mappings, without establishing any restrictions on the number of target ontologies or entities. Retrieval-Augmented Generation selects relevant classes to compose the mapping and filters matching reference mappings to serve as examples, enhancing In-Context Learning. The strategy was evaluated in three biomedical tasks with partial reference alignments. CMOMgen outperforms baselines in class selection, demonstrating the impact of having a dedicated strategy. Our strategy also achieves a minimum of 63% in F1-score, outperforming all baselines and ablated versions in two out of three tasks and placing second in the third. Furthermore, a manual evaluation of non-reference mappings showed that 46% of the mappings achieve the maximum score, further substantiating its ability to construct semantically sound mappings.


Lecture Notes on Verifying Graph Neural Networks

Schwarzentruber, François

arXiv.org Artificial Intelligence

In these lecture notes, we first recall the connection between graph neural networks, Weisfeiler-Lehman tests and logics such as first-order logic and graded modal logic. We then present a modal logic in which counting modalities appear in linear inequalities in order to solve verification tasks on graph neural networks. We describe an algorithm for the satisfiability problem of that logic. It is inspired from the tableau method of vanilla modal logic, extended with reasoning in quantifier-free fragment Boolean algebra with Presburger arithmetic.


OWL: Overcoming Window Length-Dependence in Speculative Decoding for Long-Context Inputs

Lee, Jaeseong, hwang, seung-won, Qiao, Aurick, Oliaro, Gabriele, Wang, Ye, Rajbhandari, Samyam

arXiv.org Artificial Intelligence

Speculative decoding promises faster inference for large language models (LLMs), yet existing methods fail to generalize to real-world settings. Benchmarks typically assume short contexts (e.g., 2K tokens), whereas practical workloads involve long contexts. We find current approaches degrade severely with long contexts; for instance, EAGLE3 even slows down the generation speed by 0.81x. We address these limitations by releasing a new long-context benchmark (LongSpecBench) and introducing a novel model (OWL). OWL achieves about 5x higher acceptance length than EAGLE3 on long-context inputs through three innovations: (1) an LSTM-based drafter conditioned only on the last-token state, making it generalize to various lengths, (2) a special token [SPEC] in the verifier that produces richer representation for drafter, and (3) a hybrid algorithm combining both tree and non-tree decoding methods. We release all code and datasets to advance future research.


OWL: Geometry-Aware Spatial Reasoning for Audio Large Language Models

Biswas, Subrata, Khan, Mohammad Nur Hossain, Islam, Bashima

arXiv.org Artificial Intelligence

Spatial reasoning is fundamental to auditory perception, yet current audio large language models (ALLMs) largely rely on unstructured binaural cues and single step inference. This limits both perceptual accuracy in direction and distance estimation and the capacity for interpretable reasoning. Recent work such as BAT demonstrates spatial QA with binaural audio, but its reliance on coarse categorical labels (left, right, up, down) and the absence of explicit geometric supervision constrain resolution and robustness. We introduce the $\textbf{Spatial-Acoustic Geometry Encoder (SAGE}$), a geometry-aware audio encoder that aligns binaural acoustic features with 3D spatial structure using panoramic depth images and room-impulse responses at training time, while requiring only audio at inference. Building on this representation, we present $\textbf{OWL}$, an ALLM that integrates $\textbf{SAGE}$ with a spatially grounded chain-of-thought to rationalize over direction-of-arrivals (DoA) and distance estimates. Through curriculum learning from perceptual QA to multi-step reasoning, $\textbf{OWL}$ supports o'clock-level azimuth and DoA estimation. To enable large-scale training and evaluation, we construct and release $\textbf{BiDepth}$, a dataset of over one million QA pairs combining binaural audio with panoramic depth images and room impulse responses across both in-room and out-of-room scenarios. Across two benchmark datasets, our new $\textbf{BiDepth}$ and the public SpatialSoundQA, $\textbf{OWL}$ reduces mean DoA error by $\textbf{11$^{\circ}$}$ through $\textbf{SAGE}$ and improves spatial reasoning QA accuracy by up to $\textbf{25}$\% over BAT.


Towards Understanding Subliminal Learning: When and How Hidden Biases Transfer

Schrodi, Simon, Kempf, Elias, Barez, Fazl, Brox, Thomas

arXiv.org Artificial Intelligence

Language models can transfer hidden biases during distillation. For example, a teacher that "likes owls" can make its student "like owls" too, even when the training data consists only of lists of numbers. This surprising phenomenon is called subliminal learning. Subliminal learning can be expected under soft distillation, where the student is trained on the teacher's full next-token distribution. But the fact that this also occurs under hard distillation--where the student only sees sampled tokens--raises a deeper question: when and how does subliminal learning actually occur? We answer this question through controlled experiments and mechanistic analysis. Our results show that subliminal learning does not need (global) token entanglement or logit leakage. Instead, it comes down to a small set of divergence tokens--rare cases where teachers with different biases would predict different tokens. Masking out these tokens mostly removes the hidden bias transfer. Mechanistically, divergence tokens reveal that early layers are critical. Surprisingly, finetuning even a single such early layer is sufficient for subliminal learning. Finally, we find that subliminal learning is fragile. Even small changes, like paraphrasing prompts, are usually sufficient to suppress it. Distillation is a core technique for compressing models or transferring knowledge, where a student model is trained to imitate a teacher (Hinton et al., 2015; Ba & Caruana, 2014). The common view is that what transfers depends on the (semantic) content of the training data (Dong et al., 2023; Guan et al., 2024; Chen et al., 2025; Li et al., 2025). In this view, if the teacher's outputs do not show a trait--such as a bias toward an animal or misaligned behavior--the student should not learn it.


ExeKGLib: A Platform for Machine Learning Analytics based on Knowledge Graphs

Klironomos, Antonis, Zhou, Baifan, Tan, Zhipeng, Zheng, Zhuoxun, Gad-Elrab, Mohamed H., Paulheim, Heiko, Kharlamov, Evgeny

arXiv.org Artificial Intelligence

Nowadays machine learning (ML) practitioners have access to numerous ML libraries available online. Such libraries can be used to create ML pipelines that consist of a series of steps where each step may invoke up to several ML libraries that are used for various data-driven analytical tasks. Development of high-quality ML pipelines is non-trivial; it requires training, ML expertise, and careful development of each step. At the same time, domain experts in science and engineering may not possess such ML expertise and training while they are in pressing need of ML-based analytics. In this paper, we present our ExeKGLib, a Python library enhanced with a graphical interface layer that allows users with minimal ML knowledge to build ML pipelines. This is achieved by relying on knowledge graphs that encode ML knowledge in simple terms accessible to non-ML experts. ExeKGLib also allows improving the transparency and reusability of the built ML workflows and ensures that they are executable. We show the usability and usefulness of ExeKGLib by presenting real use cases.


Explaining Multi-modal Large Language Models by Analyzing their Vision Perception

Giulivi, Loris, Boracchi, Giacomo

arXiv.org Artificial Intelligence

Multi-modal Large Language Models (MLLMs) have demonstrated remarkable capabilities in understanding and generating content across various modalities, such as images and text. However, their interpretability remains a challenge, hindering their adoption in critical applications. This research proposes a novel approach to enhance the interpretability of MLLMs by focusing on the image embedding component. We combine an open-world localization model with a MLLM, thus creating a new architecture able to simultaneously produce text and object localization outputs from the same vision embedding. The proposed architecture greatly promotes interpretability, enabling us to design a novel saliency map to explain any output token, to identify model hallucinations, and to assess model biases through semantic adversarial perturbations.


Teaching Language Models to Self-Improve through Interactive Demonstrations

Yu, Xiao, Peng, Baolin, Galley, Michel, Gao, Jianfeng, Yu, Zhou

arXiv.org Artificial Intelligence

The self-improving ability of large language models (LLMs), enabled by prompting them to analyze and revise their own outputs, has garnered significant interest in recent research. However, this ability has been shown to be absent and difficult to learn for smaller models, thus widening the performance gap between state-of-the-art LLMs and more cost-effective and faster ones. To reduce this gap, we introduce TriPosT, a training algorithm that endows smaller models with such self-improvement ability, and show that our approach can improve a LLaMA-7b's performance on math and reasoning tasks by up to 7.13%. In contrast to prior work, we achieve this by using the smaller model to interact with LLMs to collect feedback and improvements on its own generations. We then replay this experience to train the small model. Our experiments on four math and reasoning datasets show that the interactive experience of learning from and correcting its own mistakes is crucial for small models to improve their performance.


Outlier Weighed Layerwise Sparsity (OWL): A Missing Secret Sauce for Pruning LLMs to High Sparsity

Yin, Lu, Wu, You, Zhang, Zhenyu, Hsieh, Cheng-Yu, Wang, Yaqing, Jia, Yiling, Pechenizkiy, Mykola, Liang, Yi, Wang, Zhangyang, Liu, Shiwei

arXiv.org Artificial Intelligence

Large Language Models (LLMs), renowned for their remarkable performance, present a challenge due to their colossal model size when it comes to practical deployment. In response to this challenge, efforts have been directed toward the application of traditional network pruning techniques to LLMs, uncovering a massive number of parameters can be pruned in one-shot without hurting performance. Building upon insights gained from pre-LLM models, prevailing LLM pruning strategies have consistently adhered to the practice of uniformly pruning all layers at equivalent sparsity. However, this observation stands in contrast to the prevailing trends observed in the field of vision models, where non-uniform layerwise sparsity typically yields substantially improved results. To elucidate the underlying reasons for this disparity, we conduct a comprehensive analysis of the distribution of token features within LLMs. In doing so, we discover a strong correlation with the emergence of outliers, defined as features exhibiting significantly greater magnitudes compared to their counterparts in feature dimensions. Inspired by this finding, we introduce a novel LLM pruning methodology that incorporates a tailored set of non-uniform layerwise sparsity ratios specifically designed for LLM pruning, termed as Outlier Weighed Layerwise sparsity (OWL). The sparsity ratio of OWL is directly proportional to the outlier ratio observed within each layer, facilitating a more effective alignment between layerwise weight sparsity and outlier ratios. Our empirical evaluation, conducted across the LLaMA-V1 family and OPT, spanning various benchmarks, demonstrates the distinct advantages offered by OWL over previous methods. For instance, our approach exhibits a remarkable performance gain, surpassing the state-of-the-art Wanda and SparseGPT by 61.22 and 6.80 perplexity at a high sparsity level of 70%, respectively.