Goto

Collaborating Authors

 gensyn


Sharing is Caring: Efficient LM Post-Training with Collective RL Experience Sharing

Amico, Jeffrey, Andrade, Gabriel Passamani, Donaghy, John, Fielding, Ben, Forbus, Tristin, Grieve, Harry, Kara, Semih, Kolehmainen, Jari, Lou, Yihua, Nies, Christopher, Nuño, Edward Phillip Flores, Ortega, Diogo, Rastogi, Shikhar, Virts, Austin, Wright, Matthew J.

arXiv.org Artificial Intelligence

Post-training language models (LMs) with reinforcement learning (RL) can enhance their complex reasoning capabilities without supervised fine-tuning, as demonstrated by DeepSeek-R1-Zero. However, effectively utilizing RL for LMs requires significant parallelization to scale-up inference, which introduces non-trivial technical challenges (e.g. latency, memory, and reliability) alongside ever-growing financial costs. We present Swarm sAmpling Policy Optimization (SAPO), a fully decentralized and asynchronous RL post-training algorithm. SAPO is designed for decentralized networks of heterogenous compute nodes, where each node manages its own policy model(s) while "sharing" rollouts with others in the network; no explicit assumptions about latency, model homogeneity, or hardware are required and nodes can operate in silo if desired. As a result, the algorithm avoids common bottlenecks in scaling RL post-training while also allowing (and even encouraging) new possibilities. By sampling rollouts "shared" across the network, it enables "Aha moments" to propagate, thereby bootstrapping the learning process. In this paper we show SAPO achieved cumulative reward gains of up to 94% in controlled experiments. We also share insights from tests on a network with thousands of nodes contributed by Gensyn community members running the algorithm on diverse hardware and models during an open-source demo.


Machine Learning GPU Engineer at Gensyn - United Kingdom - Remote

#artificialintelligence

Supercluster: Gensyn is building a permisionless distributed network that unites all of the world's compute into a global machine learning supercluster. Supercluster: Gensyn is building a permisionless distributed network that unites all of the world's compute into a global machine learning supercluster.


GenSyn: A Multi-stage Framework for Generating Synthetic Microdata using Macro Data Sources

Acharya, Angeela, Sikdar, Siddhartha, Das, Sanmay, Rangwala, Huzefa

arXiv.org Artificial Intelligence

Individual-level data (microdata) that characterizes a population, is essential for studying many real-world problems. However, acquiring such data is not straightforward due to cost and privacy constraints, and access is often limited to aggregated data (macro data) sources. In this study, we examine synthetic data generation as a tool to extrapolate difficult-to-obtain high-resolution data by combining information from multiple easier-to-obtain lower-resolution data sources. In particular, we introduce a framework that uses a combination of univariate and multivariate frequency tables from a given target geographical location in combination with frequency tables from other auxiliary locations to generate synthetic microdata for individuals in the target location. Our method combines the estimation of a dependency graph and conditional probabilities from the target location with the use of a Gaussian copula to leverage the available information from the auxiliary locations. We perform extensive testing on two real-world datasets and demonstrate that our approach outperforms prior approaches in preserving the overall dependency structure of the data while also satisfying the constraints defined on the different variables.


Gensyn applies a token to distributed computing for AI developers, raises $6.5M – TechCrunch

#artificialintelligence

For self-driving cars and other applications developed using AI, you need what's known as'deep learning', the core concepts of which emerged in the '50s. This requires training models based on similar patterns as seen in the human brain. This, in turn, requires a large amount of compute power, as afforded by TPUs (Tensor Processing Units) or GPUs (Graphics Processing Units) running for lengthy periods. However, cost of this compute power is out of reach of most AI developers, who largely rent it from cloud computing platforms such as AWS or Azure. Well, one approach is that taken by UK startup Gensyn.