Gabriele Corso @ICLR2025 Profile picture
Oct 5, 2022 8 tweets 7 min read Read on X
Excited to share DiffDock, new non-Euclidean diffusion for molecular docking! In PDBBind, standard benchmark, DiffDock outperforms by a huge margin (38% vs 23%) the previous state-of-the-art methods that were based on expensive search!
arxiv.org/abs/2210.01776
A thread! 👇
@HannesStaerk @BarzilayRegina Recent regression-based ML methods for docking showed strong speed-up but no significant accuracy improvements over traditional search-based approaches. We identify the problem in their objective functions and show how generative modeling aligns well with the docking task.
@HannesStaerk @BarzilayRegina We, therefore, develop DiffDock, a non-Euclidean diffusion model over the space of ligand poses for molecular docking. DiffDock defines a diffusion process over the position of the ligand relative to the protein, its orientation, and the torsion angles describing its conformation
@HannesStaerk @BarzilayRegina To efficiently train and run the diffusion model over this highly non-linear manifold, we map the elements of the manifold to a product space of T(3) x SO(3) x SO(2)^m groups corresponding to the translation, rotation, and torsion transformations.
@HannesStaerk @BarzilayRegina We achieve a new state-of-the-art 38% top-1 prediction with RMSD<2A on the PDBBind blind docking benchmark, considerably surpassing the previous best search-based (23%) and deep learning methods (20%). DiffDock also has faster runtimes than the previous state-of-the-art methods.
@HannesStaerk @BarzilayRegina We also train a confidence model that can rank poses by likelihood and predict the model's confidence in the generated poses. Experimentally, this confidence score provides high selective accuracy, reaching 83% on its most confident third of the previously unseen complexes.
@HannesStaerk @BarzilayRegina The project resulted from a great collaboration with @HannesStaerk (joint first), Bowen Jing (joint first), @BarzilayRegina and Tommi Jaakkola. Code and trained models are available at: github.com/gcorso/DiffDock

• • •

Missing some Tweet in this thread? You can try to force a refresh
 

Keep Current with Gabriele Corso @ICLR2025

Gabriele Corso @ICLR2025 Profile picture

Stay in touch and get notified when new unrolls are available from this author!

Read all threads

This Thread may be Removed Anytime!

PDF

Twitter may remove this content at anytime! Save it as PDF for later use!

Try unrolling a thread yourself!

how to unroll video
  1. Follow @ThreadReaderApp to mention us!

  2. From a Twitter thread mention us with a keyword "unroll"
@threadreaderapp unroll

Practice here first or read more on our help page!

More from @GabriCorso

Apr 22
Happy to finally release our work on "Composing Unbalanced Flows for Flexible Docking and Relaxation" (FlexDock) that we will be presenting as an oral at #ICLR2025 ! 🤗✈️🇸🇬 A thread! 🧵 Image
@NoahBGetz @BarzilayRegina @arkrause TLDR: We studied the problem of flexible molecular docking and the issues with existing methods for the task. We came up with a couple of interesting technical ideas that we validated at small scale in this work and are now making their way into upcoming versions of Boltz! 🔥🚀
@NoahBGetz @BarzilayRegina @arkrause Diffusion models have emerged as a successful approach for molecular docking, but they often struggle to model protein flexibility or generate physical poses. We argue that both these challenges can be tackled by framing the problem as a transport between distributions.
Read 7 tweets
Nov 17, 2024
Thrilled to announce Boltz-1, the first open-source and commercially available model to achieve AlphaFold3-level accuracy on biomolecular structure prediction! An exciting collaboration with @jeremyWohlwend, @pas_saro and an amazing team at MIT and Genesis Therapeutics. A thread!
@jeremyWohlwend @pas_saro We test Boltz-1 on various benchmarks and show it matches the performance of Chai-1. E.g. on CASP15, Boltz-1 demonstrates strong protein-ligand and protein-protein performance achieving an LDDT-PLI of 65% (40% for Chai-1), and a proportion of DockQ>0.23 of 83% (76% for Chai-1) Image
@jeremyWohlwend @pas_saro By releasing training and inference code, model weights, and datasets under the MIT license, we aim to establish Boltz-1 as a modeling backbone for researchers worldwide, setting a new standard in open-source structural biology.
Read 6 tweets
Feb 29, 2024
Excited to finally be able to share our ICLR work critically analyzing the capacity of deep learning docking methods to generalize and how to improve this (spoiler scaling, augmentation and RL)! With this, we release a new significantly improved version of DiffDock!

A thread! 🧵 Image
Solving the general blind docking task would have profound biomedical implications. It would help us understand the mechanism of action of new drugs, predict adverse side-effects before clinical trials… But all these require methods to generalize beyond few well-studied proteins
First, we realized UniProt IDs or sequence similarity splits do not properly distinguish between evolutionarily conserved pockets. Instead, we propose DockGen a new benchmark based on binding protein domain splits and compatible with PDBBind training Image
Read 10 tweets

Did Thread Reader help you today?

Support us! We are indie developers!


This site is made by just two indie developers on a laptop doing marketing, support and development! Read more about the story.

Become a Premium Member ($3/month or $30/year) and get exclusive features!

Become Premium

Don't want to be a Premium member but still want to support us?

Make a small donation by buying us coffee ($5) or help with server cost ($10)

Donate via Paypal

Or Donate anonymously using crypto!

Ethereum

0xfe58350B80634f60Fa6Dc149a72b4DFbc17D341E copy

Bitcoin

3ATGMxNzCUFzxpMCHL5sWSt4DVtS8UqXpi copy

Thank you for your support!

Follow Us!

:(