Abstract
Abstract
Transformer models coupled with a simplified molecular line entry system (SMILES) have recently proven to be a powerful combination for solving challenges in cheminformatics. These models, however, are often developed specifically for a single application and can be very resource-intensive to train. In this work we present the Chemformer model—a Transformer-based model which can be quickly applied to both sequence-to-sequence and discriminative cheminformatics tasks. Additionally, we show that self-supervised pre-training can improve performance and significantly speed up convergence on downstream tasks. On direct synthesis and retrosynthesis prediction benchmark datasets we publish state-of-the-art results for top-1 accuracy. We also improve on existing approaches for a molecular optimisation task and show that Chemformer can optimise on multiple discriminative tasks simultaneously. Models, datasets and code will be made available after publication.
Subject
Artificial Intelligence,Human-Computer Interaction,Software
Reference55 articles.
1. Attention is all you need;Vaswani,2017
2. Long short-term memory;Hochreiter;Neural Comput.,1997
3. Learning phrase representations using RNN encoder–decoder for statistical machine translation;Cho,2014
4. Molecular transformer: a model for uncertainty-calibrated chemical reaction prediction;Schwaller;ACS Cent. Sci.,2019
5. State-of-the-art augmented NLP transformer models for direct and single-step retrosynthesis;Tetko;Nat. Commun.,2020
Cited by
124 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献