g-DPO: Scalable Preference Optimization for Protein Language Models
Ferragu, Constance, Ziegler, Jonathan D., Deutschmann, Nicolas, Lindoulsi, Arthur, Bixby, Eli, Team, Cradle ML
–arXiv.org Artificial Intelligence
Direct Preference Optimization (DPO) is an effective approach for aligning protein language models with experimental design goals. However, DPO faces a scalability bottleneck: the number of possible training pairs grows quadratically with the number of labeled sequences, leading to prohibitive training times even for modestly sized datasets. We introduce g-DPO, a framework that (i) uses sequence space clustering to prune redundant pairs while preserving training signal, and (ii) amortizes likelihood computations with group-based approximations. Across three protein engineering tasks, g-DPO maintains in silico and in vitro performance that is statistically indistinguishable from standard DPO, while converging 1.7x to 5.4x times faster, with speedups that scale with dataset size and the structure of the underlying mutational landscape.
arXiv.org Artificial Intelligence
Nov-27-2025
- Country:
- Europe
- France (0.04)
- Switzerland > Zürich
- Zürich (0.14)
- Europe
- Genre:
- Research Report
- Experimental Study (0.46)
- New Finding (0.68)
- Research Report
- Industry:
- Technology: