Now out in JACS! π : "Computing Solvation Free Energies of Small Molecules with Experimental Accuracy"! It's been a pleasure to collaborate on this with Harry Moore (@jhmchem.bsky.social) & GΓ‘bor CsΓ‘nyi pubs.acs.org/doi/10.1021/...
@weitse-hsu
- Postdoc in Drug Design at Oxford Biochemistry (Biggin Lab). - Ph.D. from the Shirts Group at CU Boulder. - Keen on compchem, deep learning & education. - Rookie runner. - Originally from Taiwan. - Check my MD tutorials: https://weitsehsu.com/
Now out in JACS! π : "Computing Solvation Free Energies of Small Molecules with Experimental Accuracy"! It's been a pleasure to collaborate on this with Harry Moore (@jhmchem.bsky.social) & GΓ‘bor CsΓ‘nyi pubs.acs.org/doi/10.1021/...
New Preprint!! We show that binding entropy can be quantitatively predicted from crystallographic ensemble models, accounting for both protein conformational entropy and solvent entropy! www.biorxiv.org/content/10.6...
π Bottom line:
With careful filtering, co-folding predictions can indeed teach ML about binding affinity.
π Read the full JCIM paper: pubs.acs.org/doi/full/10....
Work with Aniket Magarkar
@boehringerglobal.bsky.social and @philbiggin.bsky.social @ox.ac.uk
(6/6)
π SI highlights:
- AEV-PLIG beats Boltz-2 in 4 target classes in the FEP benchmark (loses 1, ties 6); both are competitive with FEP+ in some cases.
- ipLDDT & ligand pLDDT are also effective filters; pTM, PAE, PDE are not
- Boltz confidence seems to generalize better than its structure module
(5/6)
β Are co-folding predictions good enough to train scoring functions?
π Yes β with careful filtering. We see no performance difference b/w models trained on:
- experimental structures
- corresponding co-folding predictions
This holds across AEV-PLIG, EHIGN, and RF-Score.
(4/6)
β When can we trust a co-folding prediction?
π From reproducing HiQBind with Boltz-1x, a few simple heuristics are recommended high-quality cofolding augmentation:
1οΈβ£ single-chain systems
2οΈβ£ Boltz confidence > 0.9
3οΈβ£ trainβtest similarity > 60%
(3/6)
β How much can data augmentation actually improve scoring?
π Short answer: only if the added data are high-quality. Adding BindingNet v1 clearly improved performance, but v2 did notβdespite being 10x largerβdue to its substantially lower quality.
Quality beats quantity.
(2/6)
π’ Can AI-Predicted Complexes Teach Machine Learning to Compute Drug Binding Affinity?
In our recent JCIM work, we tested whether co-folding models can be used for data augmentation for training ML-based scoring functions (SFs).
We asked 3 simple but critical questions. π
(1/6)