A major limitation of current deep learning reaction prediction models is stereochemistry. It is not taken into account by graph-neural networks and a weakness of text-based prediction models, like the Molecular Transformer (doi.org/10.1021/acscen…).
How can we improve? 2/N
In this work, we take carbohydrate reactions as an example. Compared to the reactions in patents (avg. 0.4 stereocentres in product), carbohydrate contain multiple stereocentres (avg. >6 in our test set), which make reactivity predictions challenging even for human experts. 3/N
Another difficulty is the availability of good quality data. Typically, there is not enough data available to train a reaction prediction model solely on one reaction class. Inspired by #NLProc (e.g. the work @seb_ruder), we explored different transfer learning techniques. 4/N
With transfer learning, we can leverage the knowledge extracted from large general reaction data sets (e.g open-source USPTO @dan2097@nmsoftware), to train better models for the prediction of specific complex reaction (here, carbo reactions). 5/N
We explore different settings: Multi-task, where we train on the generic and specific data sets simultaneously, and sequential transfer learning, where a model trained on the generic data is adapted to the specific data in a subsequent training run. 6/N
While the first scenario ensures good performance not only on the specific but also on the generic data, the second scenario is particularly interesting because of the reduced computational cost and the fact, that the generic (potentially proprietary) data is not disclosed. 7/N
We evaluated our models in numerous ways:
- random and time-split test sets of carbohydrate reactions
- recent @J_A_C_S total syntheses
- an in-house 14-step synthesis of a lipid-linked oligosaccharide (LLO) by @Giorgio_P_ 8/N
The transfer-learned models show intriguing performance across all test sets using only small specific data sets for training.
Moreover, it is the first deep learning reaction prediction work including an experimental validation. 9/N
The methods were implemented with #OpenNMT and are straightforward to adapt for any reaction class of interest. Canicalisation was done using @RDKit_org. Code and trained models are available from github.com/rxn4chemistry/… 10/N
If you have questions, how to adapt our work to your reaction domain of interest. Feel free to reach out! 12/N
At the same time, this study is my first peer-reviewed work with the @reymondgroup. I’m very grateful to my two supervisors Prof Jean-Louis Reymond (@jrjrjlr) and Teodoro Laino (@teodorolaino)! 13/N
Excited to see our carbohydrate transformer out in @NatureComms, an awesome open-access journal!
Stay tuned, more will follow! #compchem#glycotime#AI4Chem 14/N
• • •
Missing some Tweet in this thread? You can try to
force a refresh
I had planned to add all my #LINO22 highlights chronologically to my thread but there are just too many. So, I will just cherry-pick a few here.
Starting with the inspirational talk by @ben_list on the importance of catalysis for a more sustainable future.
Related to that the "Catalysis and Green Chemistry" panel discussion with Richard Schrock, Dave MacMillan (@dmac68), Liang Feng (@LiangFeng_chem), Jiangnan Li, Carla Casadevall (@CasadevallCarla) - well done!
Dave MacMillan’s (@dmac68) excellent advice for young group leaders/scientists:
- Be passionate and work on things you are truly excited about
- Be as generous as you can be and treat people with respect
@one_know_wonho: For yield prediction, how are the data labels distributed? Does the dataset also include reactant sets where no reaction happens between them (thus zero yield)?
Yes, the yields are distributed between 0 and 100%. For the Buchwald-Hartwig reactions (science.org/doi/abs/10.112…), the dataset contains more low- than high-yielding reactions. You can find more information in iopscience.iop.org/article/10.108….
2/ The AI model (VQGAN + CLIP) generated most of the image using “enzymatic chemical reactions. green chemistry. advanced unreal engine” as input.
It’s interesting that you can recognise the lab with the blackboards, the floor and the “reactions”.
Awesome! All the video recordings of #AMLD2020 are now available on youtube. Check out the ones from the fantastic speakers we had in the #AIMolecularWorld track⬇️
We compared different RXN classification methods. 📍Using a BERT model borrowed from NLP, we matched the ground truth (Pistachio, @nmsoftware) with an accuracy of 98.2%.
We did not only visualize what was important for the class predictions by looking at the different attention weights...