Not enough data to create a plot.
Try a different view from the menu above.
Telyatnikov, Lev
TopoBenchmarkX: A Framework for Benchmarking Topological Deep Learning
Telyatnikov, Lev, Bernardez, Guillermo, Montagna, Marco, Vasylenko, Pavlo, Zamzmi, Ghada, Hajij, Mustafa, Schaub, Michael T, Miolane, Nina, Scardapane, Simone, Papamarkou, Theodore
This work introduces TopoBenchmarkX, a modular open-source library designed to standardize benchmarking and accelerate research in Topological Deep Learning (TDL). TopoBenchmarkX maps the TDL pipeline into a sequence of independent and modular components for data loading and processing, as well as model training, optimization, and evaluation. This modular organization provides flexibility for modifications and facilitates the adaptation and optimization of various TDL pipelines. A key feature of TopoBenchmarkX is that it allows for the transformation and lifting between topological domains. This enables, for example, to obtain richer data representations and more fine-grained analyses by mapping the topology and features of a graph to higher-order topological domains such as simplicial and cell complexes. The range of applicability of TopoBenchmarkX is demonstrated by benchmarking several TDL architectures for various tasks and datasets.
TopoX: A Suite of Python Packages for Machine Learning on Topological Domains
Hajij, Mustafa, Papillon, Mathilde, Frantzen, Florian, Agerberg, Jens, AlJabea, Ibrahem, Ballester, Ruben, Battiloro, Claudio, Bernárdez, Guillermo, Birdal, Tolga, Brent, Aiden, Chin, Peter, Escalera, Sergio, Fiorellino, Simone, Gardaa, Odin Hoff, Gopalakrishnan, Gurusankar, Govil, Devendra, Hoppe, Josef, Karri, Maneel Reddy, Khouja, Jude, Lecha, Manuel, Livesay, Neal, Meißner, Jan, Mukherjee, Soham, Nikitin, Alexander, Papamarkou, Theodore, Prílepok, Jaro, Ramamurthy, Karthikeyan Natesan, Rosen, Paul, Guzmán-Sáenz, Aldo, Salatiello, Alessandro, Samaga, Shreyas N., Scardapane, Simone, Schaub, Michael T., Scofano, Luca, Spinelli, Indro, Telyatnikov, Lev, Truong, Quang, Walters, Robin, Yang, Maosheng, Zaghen, Olga, Zamzmi, Ghada, Zia, Ali, Miolane, Nina
We introduce TopoX, a Python software suite that provides reliable and user-friendly building blocks for computing and machine learning on topological domains that extend graphs: hypergraphs, simplicial, cellular, path and combinatorial complexes. TopoX consists of three packages: TopoNetX facilitates constructing and computing on these domains, including working with nodes, edges and higher-order cells; TopoEmbedX provides methods to embed topological domains into vector spaces, akin to popular graph-based embedding algorithms such as node2vec; TopoModelX is built on top of PyTorch and offers a comprehensive toolbox of higher-order message passing functions for neural networks on topological domains. The extensively documented and unit-tested source code of TopoX is available under MIT license at https://github.com/pyt-team.
Hypergraph Neural Networks through the Lens of Message Passing: A Common Perspective to Homophily and Architecture Design
Telyatnikov, Lev, Bucarelli, Maria Sofia, Bernardez, Guillermo, Zaghen, Olga, Scardapane, Simone, Lio, Pietro
Most of the current hypergraph learning methodologies and benchmarking datasets in the hypergraph realm are obtained by lifting procedures from their graph analogs, leading to overshadowing specific characteristics of hypergraphs. This paper attempts to confront some pending questions in that regard: Q1 Can the concept of homophily play a crucial role in Hypergraph Neural Networks (HNNs)? Q2 Is there room for improving current HNN architectures by carefully addressing specific characteristics of higher-order networks? Q3 Do existing datasets provide a meaningful benchmark for HNNs? To address them, we first introduce a novel conceptualization of homophily in higher-order networks based on a Message Passing (MP) scheme, unifying both the analytical examination and the modeling of higher-order networks. Further, we investigate some natural, yet mostly unexplored, strategies for processing higher-order structures within HNNs such as keeping hyperedge-dependent node representations, or performing node/hyperedge stochastic samplings, leading us to the most general MP formulation up to date -MultiSet-, as well as to an original architecture design, MultiSetMixer. Finally, we conduct an extensive set of experiments that contextualize our proposals and successfully provide insights about our inquiries.
ICML 2023 Topological Deep Learning Challenge : Design and Results
Papillon, Mathilde, Hajij, Mustafa, Jenne, Helen, Mathe, Johan, Myers, Audun, Papamarkou, Theodore, Birdal, Tolga, Dey, Tamal, Doster, Tim, Emerson, Tegan, Gopalakrishnan, Gurusankar, Govil, Devendra, Guzmán-Sáenz, Aldo, Kvinge, Henry, Livesay, Neal, Mukherjee, Soham, Samaga, Shreyas N., Ramamurthy, Karthikeyan Natesan, Karri, Maneel Reddy, Rosen, Paul, Sanborn, Sophia, Walters, Robin, Agerberg, Jens, Barikbin, Sadrodin, Battiloro, Claudio, Bazhenov, Gleb, Bernardez, Guillermo, Brent, Aiden, Escalera, Sergio, Fiorellino, Simone, Gavrilev, Dmitrii, Hassanin, Mohammed, Häusner, Paul, Gardaa, Odin Hoff, Khamis, Abdelwahed, Lecha, Manuel, Magai, German, Malygina, Tatiana, Ballester, Rubén, Nadimpalli, Kalyan, Nikitin, Alexander, Rabinowitz, Abraham, Salatiello, Alessandro, Scardapane, Simone, Scofano, Luca, Singh, Suraj, Sjölund, Jens, Snopov, Pavel, Spinelli, Indro, Telyatnikov, Lev, Testa, Lucia, Yang, Maosheng, Yue, Yixiao, Zaghen, Olga, Zia, Ali, Miolane, Nina
This paper presents the computational challenge on topological deep learning that was hosted within the ICML 2023 Workshop on Topology and Geometry in Machine Learning. The competition asked participants to provide open-source implementations of topological neural networks from the literature by contributing to the python packages TopoNetX (data processing) and TopoModelX (deep learning). The challenge attracted twenty-eight qualifying submissions in its two-month duration. This paper describes the design of the challenge and summarizes its main findings.
Topological Graph Signal Compression
Bernárdez, Guillermo, Telyatnikov, Lev, Alarcón, Eduard, Cabellos-Aparicio, Albert, Barlet-Ros, Pere, Liò, Pietro
Recently emerged Topological Deep Learning (TDL) methods aim to extend current Graph Neural Networks (GNN) by naturally processing higher-order interactions, going beyond the pairwise relations and local neighborhoods defined by graph representations. In this paper we propose a novel TDL-based method for compressing signals over graphs, consisting in two main steps: first, disjoint sets of higher-order structures are inferred based on the original signal --by clustering $N$ datapoints into $K\ll N$ collections; then, a topological-inspired message passing gets a compressed representation of the signal within those multi-element sets. Our results show that our framework improves both standard GNN and feed-forward architectures in compressing temporal link-based signals from two real-word Internet Service Provider Networks' datasets --from $30\%$ up to $90\%$ better reconstruction errors across all evaluation scenarios--, suggesting that it better captures and exploits spatial and temporal correlations over the whole graph-based network structure.
From Latent Graph to Latent Topology Inference: Differentiable Cell Complex Module
Battiloro, Claudio, Spinelli, Indro, Telyatnikov, Lev, Bronstein, Michael, Scardapane, Simone, Di Lorenzo, Paolo
Latent Graph Inference (LGI) relaxed the reliance of Graph Neural Networks (GNNs) on a given graph topology by dynamically learning it. However, most of LGI methods assume to have a (noisy, incomplete, improvable, ...) input graph to rewire and can solely learn regular graph topologies. In the wake of the success of Topological Deep Learning (TDL), we study Latent Topology Inference (LTI) for learning higher-order cell complexes (with sparse and not regular topology) describing multi-way interactions between data points. To this aim, we introduce the Differentiable Cell Complex Module (DCM), a novel learnable function that computes cell probabilities in the complex to improve the downstream task. We show how to integrate DCM with cell complex message passing networks layers and train it in a end-to-end fashion, thanks to a two-step inference procedure that avoids an exhaustive search across all possible cells in the input, thus maintaining scalability. Our model is tested on several homophilic and heterophilic graph datasets and it is shown to outperform other state-of-the-art techniques, offering significant improvements especially in cases where an input graph is not provided.