Probabilistic thermal stability prediction through sparsity promoting transformer representation
Zainchkovskyy, Yevgen, Ferkinghoff-Borg, Jesper, Bennett, Anja, Egebjerg, Thomas, Lorenzen, Nikolai, Greisen, Per Jr., Hauberg, Søren, Stahlhut, Carsten
–arXiv.org Artificial Intelligence
Pre-trained protein language models have demonstrated significant applicability in different protein engineering task [1, 2]. A general usage of these pre-trained transformer models producing latent representation is to use a mean pool across residue positions to reduce the feature dimensions to further downstream tasks such as predicting bio-physical properties or other functional behaviours. In this paper we provide a two-fold contribution to machine learning (ML) driven drug design. Firstly, we demonstrate the power of sparsity by promoting penalization of pretrained transformer models to secure more robust and accurate melting temperature (Tm) prediction of single-chain variable fragments with a mean absolute error of 0.23 C. Secondly, we demonstrate the power of framing our prediction problem in a probabilistic framework. Specifically, we advocate for the need of adopting probabilistic frameworks especially in the context of ML driven drug design.
arXiv.org Artificial Intelligence
Nov-10-2022