Goto

Collaborating Authors

 automorphism




Approximately Equivariant Graph Networks

Neural Information Processing Systems

Graph neural networks (GNNs) are commonly described as being permutation equivariant with respect to node relabeling in the graph. This symmetry of GNNs is often compared to the translation equivariance of Euclidean convolution neural networks (CNNs). However, these two symmetries are fundamentally different: The translation equivariance of CNNs corresponds to symmetries of the fixed domain acting on the image signals (sometimes known as active symmetries), whereas in GNNs any permutation acts on both the graph signals and the graph domain (sometimes described as passive symmetries). In this work, we focus on the active symmetries of GNNs, by considering a learning setting where signals are supported on a fixed graph. In this case, the natural symmetries of GNNs are the automorphisms of the graph. Since real-world graphs tend to be asymmetric, we relax the notion of symmetries by formalizing approximate symmetries via graph coarsening. We present a bias-variance formula that quantifies the tradeoff between the loss in expressivity and the gain in the regularity of the learned estimator, depending on the chosen symmetry group. To illustrate our approach, we conduct extensive experiments on image inpainting, traffic flow prediction, and human pose estimation with different choices of symmetries. We show theoretically and empirically that the best generalization performance can be achieved by choosing a suitably larger group than the graph automorphism, but smaller than the permutation group.


The Underappreciated Power of Vision Models for Graph Structural Understanding

Zhao, Xinjian, Pang, Wei, Xue, Zhongkai, Jian, Xiangru, Zhang, Lei, Xu, Yaoyao, Song, Xiaozhuang, Wu, Shu, Yu, Tianshu

arXiv.org Artificial Intelligence

Graph Neural Networks operate through bottom-up message-passing, fundamentally differing from human visual perception, which intuitively captures global structures first. We investigate the underappreciated potential of vision models for graph understanding, finding they achieve performance comparable to GNNs on established benchmarks while exhibiting distinctly different learning patterns. These divergent behaviors, combined with limitations of existing benchmarks that conflate domain features with topological understanding, motivate our introduction of GraphAbstract. This benchmark evaluates models' ability to perceive global graph properties as humans do: recognizing organizational archetypes, detecting symmetry, sensing connectivity strength, and identifying critical elements. Our results reveal that vision models significantly outperform GNNs on tasks requiring holistic structural understanding and maintain generalizability across varying graph scales, while GNNs struggle with global pattern abstraction and degrade with increasing graph size. This work demonstrates that vision models possess remarkable yet underutilized capabilities for graph structural understanding, particularly for problems requiring global topological awareness and scale-invariant reasoning. These findings open new avenues to leverage this underappreciated potential for developing more effective graph foundation models for tasks dominated by holistic pattern recognition.






EEFSUVA: A New Mathematical Olympiad Benchmark

Khatibi, Nicole N, Radamovich, Daniil A., Brenner, Michael P.

arXiv.org Artificial Intelligence

Recent breakthroughs have spurred claims that large language models (LLMs) match gold medal Olympiad to graduate level proficiency on mathematics benchmarks. In this work, we examine these claims in detail and assess the extent to which current benchmarks capture genuine LLM mathematical reasoning. The composition of these benchmarks, primarily drawing from the International Mathematics Olympiad (IMO) and related competitions, may overstate models reasoning ability due to potential data contamination and a narrow focus on familiar problem types. To enable a more holistic assessment of mathematical understanding, we introduce EEFSUVA, a novel benchmark curated from under circulated regional and national Olympiads of Eastern Europe and the countries from the former Soviet Union. These contests feature problems of comparable difficulty to the IMO and are renowned for demanding nonstandard problem-solving techniques, yet their problems are far less prevalent in online corpora. Preliminary results suggest that even state-of-the-art LLMs exhibit a notable performance decline on EEFSUVA relative to other Olympiad-style benchmarks. These findings also suggest the potential importance of broader evaluation datasets for a fuller assessment of mathematical reasoning and for guiding future model development.


Optimistic Online Learning in Symmetric Cone Games

Barakat, Anas, Lin, Wayne, Lazarsfeld, John, Varvitsiotis, Antonios

arXiv.org Artificial Intelligence

Weinberger and Saul [2009]), adversarial training of quantum generative models [Dallaire-Demers and Killoran, 2018, Chakrabarti et al., 2019], and facility location optimization [Brimberg, 1995, Xue and Ye, 1997] may seem unrelated at first glance. Yet, all of them can be formulated as two-player zero-sum games where each player optimizes over a structured, convex strategy space. These strategy spaces take a diversity of forms--probability simplices, trace-one positive semidefinite (PSD) matrices, and Euclidean balls--reflecting different algebraic or geometric constraints. While this shared structure suggests the potential for unified solution methods, existing algorithms remain highly fragmented, often tailored to specific geometries in special structured problems. For instance, distance metric learning can be solved using the Frank-Wolfe algorithm [Ying and Li, 2012] or Nesterov's smoothing algorithm [Nesterov, 2007]; quantum zero-sum games can be tackled using the Matrix Multiplicative Weights Update algorithm [Jain and Watrous, 2009, Jain et al., 2022]; the celebrated Fermat-Weber facility location problem can be solved using interior point methods [Xue and Ye, 1997]. This fragmented landscape of algorithms and analyses calls for the design of broadly applicable algorithms for equilibrium learning in structured games.