Goto

Collaborating Authors

 Zavadlav, Julija


Enhancing Machine Learning Potentials through Transfer Learning across Chemical Elements

arXiv.org Artificial Intelligence

Machine Learning Potentials (MLPs) can enable simulations of ab initio accuracy at orders of magnitude lower computational cost. However, their effectiveness hinges on the availability of considerable datasets to ensure robust generalization across chemical space and thermodynamic conditions. The generation of such datasets can be labor-intensive, highlighting the need for innovative methods to train MLPs in data-scarce scenarios. Here, we introduce transfer learning of potential energy surfaces between chemically similar elements. Specifically, we leverage the trained MLP for silicon to initialize and expedite the training of an MLP for germanium. Utilizing classical force field and ab initio datasets, we demonstrate that transfer learning surpasses traditional training from scratch in force prediction, leading to more stable simulations and improved temperature transferability. These advantages become even more pronounced as the training dataset size decreases. The out-of-target property analysis shows that transfer learning leads to beneficial but sometimes adversarial effects. Our findings demonstrate that transfer learning across chemical elements is a promising technique for developing accurate and numerically stable MLPs, particularly in a data-scarce regime.


Learning Non-Local Molecular Interactions via Equivariant Local Representations and Charge Equilibration

arXiv.org Artificial Intelligence

Graph Neural Network (GNN) potentials relying on chemical locality offer near-quantum mechanical accuracy at significantly reduced computational costs. By propagating local information to distance particles, Message-passing neural networks (MPNNs) extend the locality concept to model interactions beyond their local neighborhood. Still, this locality precludes modeling long-range effects, such as charge transfer, electrostatic interactions, and dispersion effects, which are critical to adequately describe many real-world systems. In this work, we propose the Charge Equilibration Layer for Long-range Interactions (CELLI) to address the challenging modeling of non-local interactions and the high computational cost of MPNNs. This novel architecture generalizes the fourth-generation high-dimensional neural network (4GHDNN) concept, integrating the charge equilibration (Qeq) method into a model-agnostic building block for modern equivariant GNN potentials. A series of benchmarks show that CELLI can extend the strictly local Allegro architecture to model highly non-local interactions and charge transfer. Our architecture generalizes to diverse datasets and large structures, achieving an accuracy comparable to MPNNs at about twice the computational efficiency.


Predicting solvation free energies with an implicit solvent machine learning potential

arXiv.org Artificial Intelligence

Solvation free energy, and notably hydration free energy, is generally recognized as a fundamental thermodynamic quantity of interest in computational chemistry. Defined as the work done when transferring a molecule from the gas phase to the solution (water in the case of hydration free energy), it enables the computation of several key physicochemical properties of molecules, such as solubility, partition coefficients, activity coefficients, and binding free energies in solutions [1, 2]. These properties are of great importance to the pharmaceutical, environmental, and materials sciences [3-9], prompting the organization of blind prediction SAMPL challenges [10-12] with hydration free energy as one of the main targets. In addition, Mobley et al. compiled and curated a FreeSolv database of experimentally measured hydration free energies for small neutral molecules in water [13, 14]. A wide spectrum of methods is available to calculate solvation free energy, ranging from traditional approaches such as continuum solvation models [15, 16] to recent machine learning (ML) algorithms [17-26] and their combinations [27-29]. The alchemical methods with Molecular Dynamics (MD) simulations [14, 30, 31] are typically assumed to be highly accurate but computationally expensive [32, 33]. However, both the fidelity and the efficiency highly depend on the explicitly treated degrees of freedom and the employed potential energy model.


Accurate machine learning force fields via experimental and simulation data fusion

arXiv.org Artificial Intelligence

Machine Learning (ML)-based force fields are attracting ever-increasing interest due to their capacity to span spatiotemporal scales of classical interatomic potentials at quantum-level accuracy. They can be trained based on high-fidelity simulations or experiments, the former being the common case. However, both approaches are impaired by scarce and erroneous data resulting in models that either do not agree with well-known experimental observations or are under-constrained and only reproduce some properties. Here we leverage both Density Functional Theory (DFT) calculations and experimentally measured mechanical properties and lattice parameters to train an ML potential of titanium. We demonstrate that the fused data learning strategy can concurrently satisfy all target objectives, thus resulting in a molecular model of higher accuracy compared to the models trained with a single data source. The inaccuracies of DFT functionals at target experimental properties were corrected, while the investigated off-target properties remained largely unperturbed. Our approach is applicable to any material and can serve as a general strategy to obtain highly accurate ML potentials.


Scalable Bayesian Uncertainty Quantification for Neural Network Potentials: Promise and Pitfalls

arXiv.org Artificial Intelligence

Neural network (NN) potentials promise highly accurate molecular dynamics (MD) simulations within the computational complexity of classical MD force fields. However, when applied outside their training domain, NN potential predictions can be inaccurate, increasing the need for Uncertainty Quantification (UQ). Bayesian modeling provides the mathematical framework for UQ, but classical Bayesian methods based on Markov chain Monte Carlo (MCMC) are computationally intractable for NN potentials. By training graph NN potentials for coarse-grained systems of liquid water and alanine dipeptide, we demonstrate here that scalable Bayesian UQ via stochastic gradient MCMC (SG-MCMC) yields reliable uncertainty estimates for MD observables. We show that cold posteriors can reduce the required training data size and that for reliable UQ, multiple Markov chains are needed. Additionally, we find that SG-MCMC and the Deep Ensemble method achieve comparable results, despite shorter training and less hyperparameter tuning of the latter. We show that both methods can capture aleatoric and epistemic uncertainty reliably, but not systematic uncertainty, which needs to be minimized by adequate modeling to obtain accurate credible intervals for MD observables. Our results represent a step towards accurate UQ that is of vital importance for trustworthy NN potential-based MD simulations required for decision-making in practice.