Skip to main content
ACS AuthorChoice logoLink to ACS AuthorChoice
. 2023 Jun 5;63(12):3629–3636. doi: 10.1021/acs.jcim.3c00434

DrugEx: Deep Learning Models and Tools for Exploration of Drug-Like Chemical Space

Martin Šícho †,, Sohvi Luukkonen , Helle W van den Maagdenberg , Linde Schoenmaker , Olivier J M Béquignon , Gerard J P van Westen †,*
PMCID: PMC10306259  PMID: 37272707

Abstract

graphic file with name ci3c00434_0004.jpg

The discovery of novel molecules with desirable properties is a classic challenge in medicinal chemistry. With the recent advancements of machine learning, there has been a surge of de novo drug design tools. However, few resources exist that are user-friendly as well as easily customizable. In this application note, we present the new versatile open-source software package DrugEx for multiobjective reinforcement learning. This package contains the consolidated and redesigned scripts from the prior DrugEx papers including multiple generator architectures, a variety of scoring tools, and multiobjective optimization methods. It has a flexible application programming interface and can readily be used via the command line interface or the graphical user interface GenUI. The DrugEx package is publicly available at https://github.com/CDDLeiden/DrugEx.

1. Introduction

Drug discovery is a tedious and resource-intensive process that can take decades and on average costs millions of dollars.1 Computer-aided drug design facilitates this process by selecting promising compounds over ones with a poor prognosis. Using de novo drug design (DNDD), novel hit, lead, and future drug candidates can be found by exploring the vastness of the drug-like chemical space (∼1063 molecules).2

Rapid technological improvements over the last decades have led to the rising popularity of advanced machine learning methods.35 These developments have also greatly influenced the field of DNDD with state-of-the-art methods including population-based metaheuristics, recurrent neural networks (RNNs), generative adversarial networks, variational autoencoders, and transformers.68 Moreover, concepts such as transfer, conditional, and reinforcement learning (RL) are often applied to generate molecules with desired properties.

Typical objectives guiding the drug discovery process are maximization of predicted efficiencies, synthetic accessibility or drug-likeness of the compounds, and minimizing off-target effects and toxicity. Even without optimization toward favorable physicochemical and pharmacokinetic properties, DNDD is inherently a multiobjective optimization (MOO) problem.7,9

In this application note, we present the new open-source software library DrugEx, a tool for de novo design of small molecules with deep learning generative models in a multiobjective RL framework. This comprehensive tool represents the consolidation of the original work of Liu et al.’s multiple scripts based DrugEx releases. The first version of DrugEx10 consisted of an RNN single-task agent of gated recurrent units (GRU) which were updated to long short-term memory (LSTM) units in the second version,11 also introducing MOO-based RL and an updated exploitation-exploration strategy. In its third version,12 generators based on a variant of the transformer13,14 and a novel graph-based encoding allowing for the sampling of molecules with specific substructures were introduced. These developments were built on the work of Olivecrona et al.15 for the use of reinforcement learning and those of Arús-Pous et al.16 and Yang et al.’s SynthaLinker(17) for the recurrent neural network and transformer architectures, respectively.

In Section 2.1, we describe the currently available generator algorithms, the different training modes, and data preprocessing steps and amend the recently introduced graph encoding of molecules from ref (12). In Section 2.2, we present the three steps to score compounds for the RL, the computation and scaling of scores per objective, the multiobjective optimization, and detail some predefined options. Furthermore, to facilitate usage, this work is supplemented with a rich Python application programming interface (API), a command line interface (CLI), and a graphical user interface (GUI) that are described in Section 3. Finally, pretrained models are made publicly available to ease the de novo design of molecules. This application note gives an overview of all these capabilities that have been consolidated in the DrugEx package. Thereby allowing users to more easily apply de novo drug design techniques and customize them in the way they see fit.

2. Application Overview

2.1. Molecular Generator

2.1.1. Algorithms

The original DrugEx articles describe six different generator architectures.1012 The current DrugEx package includes four of these models: two SMILES-based recurrent neural networks using GRU or LSTM units and sequence- and graph-based transformers using fragments as input. The fragment-based LSTM models with and without attention from ref (12) have been discontinued as they were outperformed by the other models. The available models are briefly introduced below, and the detailed model architectures are described in Section S1.

Recurrent Neural Networks

RNNs are used to create molecules without the use of input fragments. These molecules are generated in the form of tokenized SMILES sequences. The RNNs are built from LSTM units18 or GRUs.19 The RNN model consists of the following layers: an embedding layer, three recurrent layers, a linear layer, and a softmax activation layer. These building blocks are trained to predict the most likely next output token. Compared to the transformer models, this generator does not require inputs, is quick to train, and still has a relatively low error rate.

Transformers

In addition to the token-based RNNs, DrugEx includes two fragment-based models that are variants of the transformer model using either graphs or sequences as molecular representations. For the fragment-based modeling, molecules are constructed from building blocks (detailed in Section 2.1.2). These fragments are combined to create fragmented scaffolds, which form the input for the model, and are grown into novel molecules.

Sequence-Based Transformer

The sequence-based transformer model is a decoder-only transformer that applies a multiheaded attention architecture and position-wise feed-forward layers followed by a linear layer and an activation function to predict the most likely output token.13,14 In contrast to the RNN, the transformer models allow for user-defined inputs in the form of fragments. Furthermore, contrary to the graph-based transformer, the sequence-based transformer allows for the direct incorporation of stereochemistry defined by the molecular notation used.

Graph-Based Transformer

The graph-based transformer variant deals with the positional encodings differently from the more classical sequence-based transformer encodings.13 As with a graph representation, the atom index cannot directly be used, and the encoding is a combination of the atom index (current position) and the connected atom index (previous position).12 For more details on the graph representation of the molecules, see Section 2.1.2. The graph-based model consists of a transformer encoder and a GRU-based decoder. The graph transformer has some considerable advantages compared to the sequence-based transformer as it creates only valid molecules and has a higher incorporation rate of fragments.

2.1.2. Data Preprocessing

The following section describes the default implementation of molecular preparation, i.e., standardization and fragmentation in DrugEx (detailed information in Section S2). Nevertheless, custom steps can easily be implemented with the provided Python API (Section 3.1).

In short, standardization is applied, ensuring that only organic molecules are kept. Then, for the transformer models, fragmentation is performed using the BRICS20 or RECAP21 algorithms. Combinations of the obtained fragments are made for the model to be pretrained or fine-tuned on hybridizing these fragments to form the original molecules. For the RNN, no fragmentation is performed as the model creates molecules from an empty solution. Encoding of the inputs differs between sequence- and graph-based models. Figure 1 illustrates the encoding types per generator algorithm and gives a detailed description of the graph-based encoding as it amends that described in ref (12). Details on the encoding of SMILES sequences are available in Section S2. Both the fragmentation and encoding processes determine the minimum and maximum sizes of molecules used for training which are between 200 and 1000 Da, with default parameters (Figure S3).

Figure 1.

Figure 1

Correspondence of input and encoding types with generator models. Input molecules are fragmented for sequence and graph transformers (A), and then input molecules or molecule-fragment pairs are encoded (B) before being used for training and/or sampling by the three generator architectures available (C). Graph encoding matrix of acetaminophen (D) based on the vocabulary next to it. To be encoded as a graph, the molecule is split into three fragments by the BRICS algorithm along the bonds on both sides of the nitrogen atom. Based on the atom-type and bond-type vocabulary encodings, a graph matrix is constructed. This matrix consists of five rows: (i) the current atom type as encoded by the vocabulary, (ii) the 0-based atom index in the molecule, (iii) the index of one of its neighboring atoms (if starting a fragment the index of the atom itself), (iv) the bond type as encoded by the vocabulary, and (v) the 1-based index of the fragment being encoded, respectively. The matrix consists of four major column blocks, from left to right, the start token block (⟨GO⟩, pink), the columns used for the encoding of fragments (green), the end token block (⟨EOS⟩, pink), and columns indicating the linking between fragments (purple). The dimension of the graph matrix is Inline graphic with Inline graphic, where nfragments is the number of fragments encoded in the molecule and d the width of the block encoding fragments. Should d be greater than the number of columns required to encode all fragments of a molecule, the remaining columns are filled with zeros, as exemplified by the sub-block used for padding. During sampling, this sub-block is used to grow the molecule. By default, the graph matrix has dimensions 5 × 400 theoretically allowing for the encoding of molecules with molecular weights of up to 10,000 g/mol.

2.1.3. Training

Pretraining and Transfer Learning

Before guiding a generator to create compounds with specific properties, it needs to be pretrained to learn the language of drug-like molecules and be able to generate reasonable molecules. This involves training a generator to reproduce compounds from a large set of (fragmented) drug-like molecules. We have shared pretrained RNN-based and transformer-based generators (pretrained on ChEMBL27, ChEMBL31,22 and Papyrus v5.523) on Zenodo.24 Further details about the pretrained models are given in Sections S1 and S4.

Furthermore, a generator can be directed toward the desired chemical space by fine-tuning a pretrained generator via transfer learning with a set of molecules occupying the desired chemical space.

During training, the loss on a separate test set is assessed at each training epoch to select the best model epoch and allow for early stopping. In brief, for all models, the loss is calculated by taking the average negative log-likelihood (softmax) of the predicted outputs. For the SMILES-based model, it is also possible to use SMILES validity for this purpose.

Reinforcement Learning

During reinforcement learning, the “desirability” of generated molecules is quantified by the environment based on various properties and used as a reward (Section 2.2). The generator is optimized using the policy gradient scheme.25

In order to control the exploration rate, molecules are generated based on the output of two generators: a generator that is updated based on the reward function and saved as the final generator (the exploitation network or “agent”) and a static generator (the exploration network or “prior”). The fraction of outputs coming from each generator mimics the mutation rate in evolutionary algorithms and is tunable. Currently, this has been tested with the fine-tuned model as the exploitation network and the pretrained model as the exploration network.1012 To improve exploration, the exploitation network can use the outputs from two networks, of which one is constantly updated based on the reward function and the other only updated every 50 epochs by default (as is shown by Figure 3 in the original paper11). This is the default for the RNN-based generator. For the transformer-based generators, due to their higher computation costs, this periodically updated generator is not used by default, and we generally advise against using it with them. A reload interval of 50 epochs seems to provide a good balance between the computational cost of reloading and added benefit of more exploration. However, we have yet to explore the effect of this parameter fully. Multiple metrics are computed at each epoch: the ratio of valid, accurate (only for fragment-based models), unique, or desired molecules, and the average arithmetic and geometric mean score per objective. In addition to this, users can define customized metrics in the API. A compound is “desired” if it fulfils all objectives as defined in Section 2.2.2. One of the metrics, by default the desirability ratio, is used to select the best model epoch and to allow for early stopping. For fragment-based sampling, the inputs can either be a specific scaffold or the unique fragment combinations of a given data set.

2.2. Molecule Scoring

The scoring of molecules with the environment at each reinforcement learning epoch is done in three stages: (i) obtaining raw scores for each of the selected objectives (Section 2.2.1), (ii) scaling of the raw scores with modifier functions (Section 2.2.2), and (iii) a multiobjective optimization step (Section 2.2.3) to obtain a final reward per molecule.

2.2.1. Objectives

The API gives a large flexibility to the user to use custom scoring methods that take SMILES as inputs and give a numeric score as output. Moreover, DrugEx is coupled with the QSPRpred package to allow the use of a wide range of ML models and offers a range of other predefined objective functions.

QSPRpred

To optimize the binding affinity for one or more targets or other molecular properties, DrugEx users can choose to add one or more quantitative-structure activity/property (QSAR/QSPR) models as objectives for the reinforcement learning reward. To this end, DrugEx is compatible with any Python script that receives SMILES as input and produces a score as an output through the API. A separate package, QSPRpred (https://github.com/CDDLeiden/QSPRPred), was developed to simplify the development of QSAR models. The setup of QSPRpred is very similar to DrugEx, using the same structure of the API and CLI. It also comes with tutorials to help users get started. QSPRpred has a selection of scikit-learn26 models and a PyTorch27 fully connected neural network available through the API, so the user can train a wide variety of QSAR models. Furthermore, due to its modularity, QSPRpred is customizable; users can for example add new model types and molecular descriptors.

Predefined Objectives

The DrugEx package offers a set of predefined property calculations that can be used as objectives in the scoring environment. These components are summarized in Section S3 and include functions to compute ligand or lipophilic efficiencies from affinity predictions, a variety of similarity measures to a reference structure, estimations of (retro)synthetic accessibility, and a plethora of physicochemical descriptors.

2.2.2. Modifiers

To ensure the optimization, each objective is coupled with a modifier function that transforms it into a maximization task and scales all raw scores between 0 and 1. Custom modifiers are easily implemented with the API, but DrugEx offers a variety of predefined modifiers for both monotonic (e.g., ClippedScore) and nonmonotonic objectives (e.g., Gaussian). Some of these modifiers do not normalize or do not transform the objectives to maximization tasks in all cases and should be used with caution, especially when using an aggregation-based multiobjective optimization scheme. All modifiers are summarized in Section S3. Each objective–modifier couple is associated with a desirability threshold set between 0 and 1 to determine if a compound fulfills the desirability criteria on that objective or not. A compound is considered desired if for all objectives its modifier scores are above their corresponding thresholds.

2.2.3. Multiobjective Optimization

Since version 2, DrugEx enables multiobjective optimization during RL.11 DrugEx offers three different MOO schemes: a parametric aggregation method and two Pareto ranking-based schemes.7,9 The aggregation method is the parametric weighted sum (WS) which uses dynamic weights for each objective to especially reward compounds that perform well on the worst-performing objective(s) at each iteration.11 Pareto-based schemes do not combine multiple objectives into one but rather search for the best trade-off between them, and the initial ranking of molecules is done based on the ranking of the Pareto frontiers. After assigning each molecule to a front, the compounds in each frontier are ranked based on a distance metric to increase the diversity of solutions.

DrugEx proposes two distance metric formulations: the crowding distance (PRCD)28 and the Tanimoto distance (PRTD).11 Crowding distance is commonly used in multiobjective optimization tasks and is designed to increase the diversity of solutions with regards to how they fulfill each objective and, thus, reflects the diversity of solutions purely by their position in the objective space. In the context of molecular generation, this can sometimes be at the expense of molecular diversity. Therefore, several subtly different ranking schemes were devised utilizing the Tanimoto distance, which relates to the structural diversity of the molecules in the front. Detailed descriptions of the three schemes are given in Section S3.

The choice of an appropriate optimization scheme depends on the task at hand and usually requires some level of experimentation. The WS is more stable for many-objective optimization than the Pareto-based methods as the number of Pareto-equivalent solutions increases with the number of objectives. On the other hand, the Pareto-based schemes enforce diversity in the ranking which is not the case for the WS. The computational complexity of the WS is lower than the Pareto’s and is much faster. However, the ranking computation time (below 1 s for 10,000 molecules with three objectives with both approaches) is negligible compared to the other steps of the reinforcement learning training (generation, decoding, scoring).

3. Implementation

Aside from the addition of several new features for the generation and scoring of molecular structures, a significant part of the development was dedicated to creating a flexible and scalable software architecture. We have extensively revised the original Python source code of all published DrugEx models1012 and transformed it into a self-contained open-source Python package with a clear structure and API. In addition, a simple CLI was also implemented that allows quick invocation of the main DrugEx functions and improves the management of inputs and outputs. The package supports many monitoring utilities that log training progress and result in easy-to-read machine-readable formats such as TSV (Tab Separated Values) and JSON (JavaScript Object Notation) files. When using the CLI, these files are backed up after each (even unsuccessful) run so older results and settings are not lost and can be retrieved at any time. These and other modifications should empower users to quickly explore different scenarios when building their generative models and also ensure reproducible results by keeping track of the set parameters. Both the CLI and Python API are documented, and we also created easy-to-follow Jupyter notebooks tutorials to help users get started. Finally, we have performed significant optimizations in multiple parts of the workflow by utilizing multiprocessing where possible.

3.1. Python Package

The software package is divided into intuitively organized subpackages and modules, each handling either preprocessing of the data, model training, or generation of new molecules by loading the model for sampling (Figure 2). The code is organized with modularity, extensibility, and testability in mind. Each larger subpackage contains a clear definition of its interfaces in the interfaces.py module and also unit tests in tests.py. Interface definitions are most of the time facilitated through abstract classes for which each subpackage provides default implementations that can be automatically tested with the provided unit tests. These classes form the core of the DrugEx Python API that users can exploit to make modifications to their workflows and/or interact with the DrugEx models programmatically.

Figure 2.

Figure 2

A simplified diagram of the open-source Python package architecture. The two main subpackages of the drugex package are data and training. The data package handles the preparation of data sets that can then be used to instantiate and train models in training. Aside from the model classes themselves (generators), the training package also contains data structures needed to train models with reinforcement learning (explorers) using a scoring environment based on scoring functions (scorers).

Application Programming Interface

The Python API exposes many functions from preprocessing, model training, and sampling of molecular structures. Users can mix and match the necessary objects or create custom classes to accomplish their goals. For example, it is possible to apply customized fragmentation strategies by implementing the fragmentation API or use a modified training monitor class to change progress and result tracking during model training. All components of the project follow the principles of object-oriented programming to make implementations of such extensions possible without changing the code of the package itself. Examples of such usage are also available in the tutorials (see Tutorials).

Command Line Interface

If no customization is required, the package also offers a CLI for quick setup of experiments with default implementations of the most common tasks. The package contains three main executable scripts: (1) dataset.py, (2) train.py, and (3) generate.py. These scripts are usually executed in order to preprocess input data, train new models, and generate a virtual library of compounds. These scripts are installed with the package drugex. Each script also automatically logs standard input and output, tracks the history of executed commands, and stores generated data outputs so that they can be retrieved later which adds to the reproducibility of experiments.

Documentation

Both API and CLI usage is documented, and we have tried to provide a sufficient description of each interface, class, and function. This Sphinx-generated documentation is available at https://cddleiden.github.io/DrugEx/docs/ and is updated with each new DrugEx release.

Tutorials

Aside from source code documentation, the DrugEx web page also provides descriptions of command line arguments and usage examples for the CLI. In addition, we also compiled a collection of Jupyter notebooks that provides a comprehensive introduction to the Python API. The tutorials feature more advanced concepts and are a good starting point for any users who require more customization or any future contributors to familiarize themselves with the code.

Hardware Requirements

DrugEx offers models of different complexities, and thus, the hardware requirements vary with each model. To be able to train and use all models in the package, the user needs at least one GPU compatible with CUDA 9.2 and at least 8 GB of video memory to save the model and sufficiently large training batches. However, the basic sequential RNN model should be possible to fine-tune and optimize with reinforcement learning even on a less optimal configuration. For the two transformers, we recommend using multiple GPUs to increase throughput by parallelization, which is automatically handled by the package. The GPUs used herein are detailed in the caption of Table S3 in Section S4

Contributions to the Open-Source Code

The project embraces the open-source philosophy and welcomes all contributions, questions, or feature requests. The project’s GitHub page contains all the necessary information for potential contributors. In summary, users are advised to first initiate a community discussion on the project’s public issue tracker on GitHub. Subsequently, when an agreement is reached, the contribution can be made through a pull request.

3.2. Graphical User Interface

We also added support for the new DrugEx features to our GenUI platform,29 which provides a graphical user interface for molecular generators. GenUI is an open-source web-based application built with the Django web framework30 and the RDKit cheminformatics toolkit.31 GenUI provides features for easy integration of cheminformatics tasks commonly used in de novo generation of molecules (i.e., management of a compound database, QSAR modeling, and chemical space visualization). As of now, most of the features available through the DrugEx Python package are also exposed in this GUI to allow quick creation and management of generative workflows from the import of the training data to interactive visual analysis of the generated and real chemical space. One notable feature of the new GUI is the interactive creation of scoring environments, which makes the setup of desirability modifier functions for the multiobjective optimization more intuitive (Figure 3). GenUI is available via prebuilt Docker images and, thus, can be readily deployed in-house or online in virtual environments. Its scalable client-worker architecture also makes it easy to add or remove computation nodes as needed even while the application is running.

Figure 3.

Figure 3

An impression of GenUI interactive interface for visualizing and creating desirability modifiers. The shown view is displayed after the user selects a modifier function and a range of values for its visualization. The modifier function parameters are then simply adjusted with sliders or inputting parameter values directly into input fields (not shown). In the shown example, a Smooth Hump function is used to give maximum score to structures with molecular weight between 250 and 400 Da.

4. Conclusion

In this paper, we have described the DrugEx open-source software package that facilitates the training of a diverse set of generative models for de novo design of small molecules. The package is based on the original Python scripts previously introduced by Liu et al. that were used to develop and validate these models.1012 It includes the following new features: early stopping in all training modes, additional predefined scoring functions, and improved QSPR modeling (hyperparameter optimization, new input features, etc.) with the separate QSPRpred package. The performance has also been enhanced by utilizing parallel processing where possible in both the DrugEx and QSPRpred packages. Furthermore, the current implementation features major revisions of the original API source code of which most notable are the addition of a CLI and Python API. A GUI is also provided via the GenUI web application.

We envisage that the new DrugEx software package and its GenUI integration should be suitable for a diverse set of users. On one side, the package provides a quick and easy way to set up experiments and build models via the CLI and GUI, but on the other side, it also enables more advanced alterations to the workflow through the new Python API. The documentation was also significantly improved, and we now provide easy-to-follow tutorials for new users. Finally, all software presented in this work is provided as open-source software and accessible at https://github.com/CDDLeiden/DrugEx.

We regard the publication of this package as an important step in the development of DrugEx that will be the basis for many research projects and innovations yet to come. In fact, we believe that groundbreaking approaches are only possible when developers of generative models for chemistry undertake such open-source software development initiatives to facilitate prospective validation and testing of their new methods and most importantly their application. Additionally, providing rich documentation and tutorials helps to enhance the models’ usability and integration potential, allowing for faster adoption and feedback leading to the development of better AI-powered models and tools.

In future developments of the DrugEx package, we will not only focus on the integration of novel objectives from the drug discovery toolbox (i.e., molecular docking or retrosynthesis prediction), but also on increasing the range of possible inputs to alternative linear representations of compounds (i.e., SELFIES32) in sequence-based models or adding support for encoding stereochemistry. Moreover, we would like to focus on the development of user-centric features such as providing an even better learning platform for teaching the underlying concepts of AI-based molecular generation and improving the GenUI integration. The potential of artificial intelligence in drug discovery is tremendous, but integrating these novel tools in current workflows still remains a challenge. We hope that our software package will help to overcome at least some of those challenges.

Acknowledgments

The authors thank Xuhan Liu, the author of the original idea to develop the DrugEx models and code; we are happy for his continuous support of the project. The authors also thank Roelof van der Kleij for his help using the university IT infrastructure, our Master student Yorick van Aalst for testing of the code, and Alan K. Hassen and Andrius Bernatavicius for fruitful discussions. Some of the used computational resources were provided by the e-INFRA CZ project (ID: 90140), supported by the Ministry of Education, Youth and Sports of the Czech Republic.

Data Availability Statement

The DrugEx software package is accessible at https://github.com/CDDLeiden/DrugEx. Pretrained models are available on Zenodo.24 Timings of pretrained models are on Zenodo as well.33

Supporting Information Available

The Supporting Information is available free of charge at https://pubs.acs.org/doi/10.1021/acs.jcim.3c00434.

  • S1: Generator architecture specification. S2: Data preprocessing. S3: Environment specifications. S4: Sampling statistics. (PDF)

Author Contributions

M.Š., S.L., H.W.v.d.M., L.S., and O.J.M.B. have contributed equally to this work. M.Š., S.L., H.W.v.d.M., L.S., and O.J.M.B: Methodology, Software, Data Curation, Writing (Original Draft & Editing) and Visualization. M.Š. and G.J.P.v.W: Funding acquisition and Resources. G.J.P.v.W: Supervision and Writing (Review). M.Š.: Methodology, Software, Data Curation, Writing - Original Draft, Writing - Review & Editing, Visualization. S.L.: Methodology, Software, Data Curation, Writing - Original Draft, Writing - Review & Editing, Visualization, Project administration. H.W.v.d.M.: Methodology, Software, Data Curation, Writing - Original Draft, Writing - Review & Editing, Visualization. L.S.: Methodology, Software, Data Curation, Writing - Original Draft, Writing - Review & Editing, Visualization. O.J.M.B.: Methodology, Software, Data Curation, Writing - Original Draft, Writing - Review & Editing, Visualization. G.J.P.v.W.: Resources, Writing - Review & Editing, Supervision, Funding acquisition.

S.L. received funding from the Dutch Research Council (NWO) in the framework of the Science PPP Fund for the top sectors and acknowledges the Dutch Research Council (NWO ENPPS.LIFT.019.010). M.Š. was supported by Czech Science Foundation Grant No. 22-17367O and by the Ministry of Education, Youth and Sports of the Czech Republic (project number LM2023052).

The authors declare no competing financial interest.

Supplementary Material

ci3c00434_si_001.pdf (7.1MB, pdf)

References

  1. Wouters O. J.; McKee M.; Luyten J. Estimated Research and Development Investment Needed to Bring a New Medicine to Market, 2009–2018. Journal of the American Medical Association 2020, 323, 844–853. 10.1001/jama.2020.1166. [DOI] [PMC free article] [PubMed] [Google Scholar]
  2. Kirkpatrick P.; Ellis C. Chemical space. Nature 2004, 432, 823. 10.1038/432823a. [DOI] [Google Scholar]
  3. Anstine D. M.; Isayev O. Generative Models as an Emerging Paradigm in the Chemical Sciences. Journal of American Chemical Society 2023, 145, 8736–8750. 10.1021/jacs.2c13467. [DOI] [PMC free article] [PubMed] [Google Scholar]
  4. Perron Q.; Mirguet O.; Tajmouati H.; Skiredj A.; Rojas A.; Gohier A.; Ducrot P.; Bourguignon M.-P.; Sansilvestri-Morel P.; Do Huu N.; Gellibert F.; Gaston-Mathé Y. Deep generative models for ligand-based de novo design applied to multi-parametric optimization. Journal of Compututational Chemistry 2022, 43, 692–703. 10.1002/jcc.26826. [DOI] [PubMed] [Google Scholar]
  5. Grisoni F. Chemical language models for de novo drug design: Challenges and opportunities. Current Opinion in Structural Biology-. 2023, 79, 102527. 10.1016/j.sbi.2023.102527. [DOI] [PubMed] [Google Scholar]
  6. Liu X.; IJzerman A. P.; van Westen G. J. P. In Artificial Neural Networks; Methods in Molecular Biology series; Cartwright H., Ed.; Springer US: New York, 2021; Vol. 2190; pp 139–165. 10.1007/978-1-0716-0826-5_6. [DOI] [PubMed] [Google Scholar]
  7. Luukkonen S.; van den Maagdenberg H. W.; Emmerich M. T.; van Westen G. J. Artificial Intelligence in Multi-objective Drug Design. Currurent Opinion in Structural Biology 2023, 79, 102537. 10.1016/j.sbi.2023.102537. [DOI] [PubMed] [Google Scholar]
  8. Mouchlis V. D.; Afantitis A.; Serra A.; Fratello M.; Papadiamantis A. G.; Aidinis V.; Lynch I.; Greco D.; Melagraki G. Advances in De Novo Drug Design: From Conventional to Machine Learning Methods. International Journal of Molecular Sciences 2021, 22, 1676. 10.3390/ijms22041676. [DOI] [PMC free article] [PubMed] [Google Scholar]
  9. Fromer J. C.; Coley C. W. Computer-aided multi-objective optimization in small molecule discovery. Patterns 2023, 4, 100678. 10.1016/j.patter.2023.100678. [DOI] [PMC free article] [PubMed] [Google Scholar]
  10. Liu X.; Ye K.; van Vlijmen H. W. T.; IJzerman A. P.; van Westen G. J. P. An exploration strategy improves the diversity of de novo ligands using deep reinforcement learning: a case for the adenosine A2A receptor. Journal of Cheminformatics 2019, 11, 35. 10.1186/s13321-019-0355-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
  11. Liu X.; Ye K.; van Vlijmen H. W. T.; Emmerich M. T. M.; IJzerman A. P.; van Westen G. J. P. DrugEx v2: de novo design of drug molecules by Pareto-based multi-objective reinforcement learning in polypharmacology. Journal of Cheminformatics 2021, 13, 85. 10.1186/s13321-021-00561-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  12. Liu X.; Ye K.; van Vlijmen H. W. T.; IJzerman A. P.; van Westen G. J. P. DrugEx v3: scaffold-constrained drug design with graph transformer-based reinforcement learning. Journal of Cheminformatics 2023, 15, 24. 10.1186/s13321-023-00694-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
  13. Vaswani A.; Shazeer N.; Parmar N.; Uszkoreit J.; Jones L.; Gomez A. N.; Kaiser L.; Polosukhin I.. Attention is All You Need. Advances in Neural Information Processing Systems 30, 2017; pp 6000–6010.
  14. Radford A.; Narasimhan K.; Salimans T.; Sutskever I.. et al. Improving language understanding by generative pre-training. Preprint, 2018.
  15. Olivecrona M.; Blaschke T.; Engkvist O.; Chen H. Molecular de-novo design through deep reinforcement learning. Journal of Cheminformatics 2017, 9, 48. 10.1186/s13321-017-0235-x. [DOI] [PMC free article] [PubMed] [Google Scholar]
  16. Arús-Pous J.; Patronov A.; Bjerrum E. J.; Tyrchan C.; Reymond J.-L.; Chen H.; Engkvist O. SMILES-based deep generative scaffold decorator for de-novo drug design. Journal of Cheminformatics 2020, 12, 38. 10.1186/s13321-020-00441-8. [DOI] [PMC free article] [PubMed] [Google Scholar]
  17. Yang Y.; Zheng S.; Su S.; Zhao C.; Xu J.; Chen H. SyntaLinker: automatic fragment linking with deep conditional transformer neural networks. Chemical Science 2020, 11, 8312–8322. 10.1039/D0SC03126G. [DOI] [PMC free article] [PubMed] [Google Scholar]
  18. Hochreiter S.; Schmidhuber J. Long Short-Term Memory. Neural Computation 1997, 9, 1735–1780. 10.1162/neco.1997.9.8.1735. [DOI] [PubMed] [Google Scholar]
  19. Cho K.; van Merriënboer B.; Bahdanau D.; Bengio Y.. On the Properties of Neural Machine Translation: Encoder–Decoder Approaches. Proceedings of SSST-8, Eighth Workshop on Syntax, Semantics and Structure in Statistical Translation, Doha, Qatar, 2014; pp 103–111. 10.3115/v1/W14-4012. [DOI]
  20. Degen J.; Wegscheid-Gerlach C.; Zaliani A.; Rarey M. On the Art of Compiling and Using ‘Drug-Like’ Chemical Fragment Spaces. ChemMedChem. 2008, 3, 1503–1507. 10.1002/cmdc.200800178. [DOI] [PubMed] [Google Scholar]
  21. Lewell X. Q.; Judd D. B.; Watson S. P.; Hann M. M. Journal of Chemical Information & Computer Science 1998, 38, 511–522. 10.1021/ci970429i. [DOI] [PubMed] [Google Scholar]
  22. Gaulton A.; Hersey A.; Nowotka M.; Bento A. P.; Chambers J.; Mendez D.; Mutowo P.; Atkinson F.; Bellis L. J.; Cibrián-Uhalte E.; Davies M.; Dedman N.; Karlsson A.; Magariños M. P.; Overington J. P.; Papadatos G.; Smit I.; Leach A. R. The ChEMBL database in 2017. Nucleic Acids Research 2017, 45, D945–D954. 10.1093/nar/gkw1074. [DOI] [PMC free article] [PubMed] [Google Scholar]
  23. Béquignon O. J. M.; Bongers B. J.; Jespers W.; IJzerman A. P.; van der Water B.; van Westen G. J. P. Papyrus: a large-scale curated dataset aimed at bioactivity predictions. Journal of Cheminformatics 2023, 15, 3. 10.1186/s13321-022-00672-x. [DOI] [PMC free article] [PubMed] [Google Scholar]
  24. a Béquignon O. J. M.DrugEx RNN-GRU pretrained model (ChEMBL31), 2023. 10.5281/zenodo.7550739. [DOI]; b Béquignon O. J. M.DrugEx RNN-GRU pretrained model (Papyrus 05.5), 2023. 10.5281/zenodo.7550792. [DOI]; c Liu X.DrugEx v2 pretrained model (ChEMBL27), 2022. 10.5281/zenodo.7096837. [DOI]; d Béquignon O. J. M.DrugEx v2 pretrained model (ChEMBL31), 2022. 10.5281/zenodo.7378916. [DOI]; e Schoenmaker L.; Béquignon O. J. M.. DrugEx v2 pretrained model (Papyrus 05.5), 2022. 10.5281/zenodo.7378923. [DOI]; f Šícho M.DrugEx pretrained model (SMILES-based; Papyrus 05.5), 2023. 10.5281/zenodo.7635064. [DOI]; g Béquignon O. J. M.DrugEx pretrained model (SMILES-based; RECAP; Papyrus 05.5), 2023. 10.5281/zenodo.7622774. [DOI]; h Liu X.DrugEx v3 pretrained model (graph-based; ChEMBL27), 2022. 10.5281/zenodo.7096823. [DOI]; i Béquignon O. J. M.DrugEx v3 pretrained model (graph-based; Papyrus 05.5), 2022. 10.5281/zenodo.7085421. [DOI]; j Béquignon O. J. M.DrugEx pretrained model (graph-based; RECAP; Papyrus 05.5), 2023. 10.5281/zenodo.7622738. [DOI]
  25. Sutton R. S.; McAllester D.; Singh S.; Mansour Y.. Policy Gradient Methods for Reinforcement Learning with Function Approximation. Advances in Neural Information Processing Systems 12, 1999.
  26. Pedregosa F.; Varoquaux G.; Gramfort A.; Michel V.; Thirion B.; Grisel O.; Blondel M.; Prettenhofer P.; Weiss R.; Dubourg V.; Vanderplas J.; Passos A.; Cournapeau D.; Brucher M.; Perrot M.; Duchesnay E. Scikit-learn: Machine Learning in Python. Journal of Machine Learning Research 2011, 12, 2825–2830. [Google Scholar]
  27. Paszke A.; Gross S.; Massa F.; Lerer A.; Bradbury J.; Chanan G.; Killeen T.; Lin Z.; Gimelshein N.; Antiga L.; Desmaison A.; Kopf A.; Yang E.; DeVito Z.; Raison M.; Tejani A.; Chilamkurthy S.; Steiner B.; Fang L.; Bai J.; Chintala S.. Advances in Neural Information Processing Systems 32; Curran Associates, Inc., 2019; pp 8024–8035. [Google Scholar]
  28. Deb K.; Jain H. An Evolutionary Many-Objective Optimization Algorithm Using Reference-Point-Based Nondominated Sorting Approach, Part I: Solving Problems With Box Constraints. IEEE Transactions on Evolutionary Computation 2014, 18, 577–601. 10.1109/TEVC.2013.2281535. [DOI] [Google Scholar]
  29. Šícho M.; Liu X.; Svozil D.; van Westen G. J. P. GenUI: interactive and extensible open source software platform for de novo molecular generation and cheminformatics. Journal of Cheminformatics 2021, 13, 73. 10.1186/s13321-021-00550-y. [DOI] [PMC free article] [PubMed] [Google Scholar]
  30. Django Web Framework. Django Software Foundation. https://djangoproject.com.
  31. RDKit: Open-source cheminformatics. RDKit. https://www.rdkit.org.
  32. Krenn M.; Häse F.; Nigam A.; Friederich P.; Aspuru-Guzik A. Self-referencing embedded strings (SELFIES): A 100% robust molecular string representation. Machine Learning: Science and Technology 2020, 1, 045024. 10.1088/2632-2153/aba947. [DOI] [Google Scholar]
  33. Šícho M.; Luukkonen S.; van den Maagdenberg H. W.; Schoenmaker L.; Béquignon O. J. M.; van Westen G. J. P.. Sampling timings of DrugEx pretrained models, 2023. 10.5281/zenodo.7928362. [DOI] [PMC free article] [PubMed]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

ci3c00434_si_001.pdf (7.1MB, pdf)

Data Availability Statement

The DrugEx software package is accessible at https://github.com/CDDLeiden/DrugEx. Pretrained models are available on Zenodo.24 Timings of pretrained models are on Zenodo as well.33


Articles from Journal of Chemical Information and Modeling are provided here courtesy of American Chemical Society

RESOURCES