Abstract
During the drug design process, one must develop a molecule, which structure satisfies a number of physicochemical properties. To improve this process, we introduce Mol-CycleGAN – a CycleGAN-based model that generates compounds optimized for a selected property, while aiming to retain the already optimized ones. In the task of constrained optimization of penalized logP of drug-like molecules our model significantly outperforms previous results.
You have full access to this open access chapter, Download conference paper PDF
Similar content being viewed by others
Keywords
1 Introduction
The principal goal of the drug design process is to find new chemical compounds that are able to modulate the activity of a given target in a desired way [13]. However, finding such molecules in the high-dimensional chemical space of all molecules without any prior knowledge is nearly impossible. In silico methods have been introduced to leverage the existing knowledge, thus forming a new branch of science - computer-aided drug design (CADD) [1, 12].
The recent advancements in deep learning have encouraged its application in CADD [4]. One of the main approaches is de novo design, that is using generative models to propose new molecules that are likely to possess the desired properties [3, 5, 15, 17].
In the center of our interest are the hit-to-lead and lead optimization phases of the compound design process. Their goal is to optimize the drug candidates identified in the previous steps in terms of the desired activity profile and their physicochemical and pharmacokinetic properties.
To address this problem, we introduce Mol-CycleGAN – a generative model based on CycleGAN [19]. Given a starting molecule, it generates a structurally similar one but with a desired characteristic. We show that our model generates molecules that possess desired properties while retaining their structural similarity to the starting compound. Moreover, thanks to employing graph-based representation, our algorithm always returns valid compounds.
To assess the model’s utility for compound design we evaluate its ability to maximize penalized logP property. Penalized logP is chosen because it is often selected as a testing ground for molecule optimization models [7, 18], due to its relevance in the drug design process. In the optimization of penalized logP for drug-like molecules our model significantly outperforms previous results. To the best of our knowledge, Mol-CycleGAN is the first approach to molecule generation that uses the CycleGAN architecture.
2 Mol-CycleGAN
Mol-CycleGAN is a novel method of performing compound optimization by learning from the sets of molecules with and without the desired molecular property (denoted by the sets X and Y). Our approach is to train a model to perform the transformation \(G: X \rightarrow Y\) (and \(F: Y \rightarrow X\)) which returns the optimized molecules. In the context of compound design X (Y) can be, e.g., the set of inactive (active) molecules.
To represent the sets X and Y our approach requires an embedding of molecules which is reversible, i.e. enables both encoding and decoding of molecules. For this purpose we use the latent space of Junction Tree Variational Autoencoder (JT-VAE) [7] – we represent each molecule as a point in the latent space, given by the mean of the variational encoding distribution [9]. This approach has the advantage that the distance between molecules (required to calculate the loss function) can be defined directly in the latent space.
Our model works as follows: (i) we define the sets X and Y (e.g., inactive/active molecules); (ii) we introduce the mapping functions \(G: X \rightarrow Y\) and \(F: Y \rightarrow X\); (iii) we introduce discriminator \(D_X\) (and \(D_Y\)) which forces the generator F (and G) to generate samples from a distribution close to the distribution of X (or Y). The components F, G, \(D_X\), and \(D_Y\) are modeled by neural networks (see subsect. 2.1 for technical details).
The main idea is to: (i) take the prior molecule x without a specified feature (e.g. activity) from set X, and compute its latent space embedding; (ii) use the generative neural network G to obtain the embedding of molecule G(x), that has this feature (as if the G(x) molecule came from set Y) but is also similar to the original molecule x; (iii) decode the latent space coordinates given by G(x) to obtain the optimized molecule. Thereby, the method is applicable in lead optimization processes, as the generated compound G(x) remains structurally similar to the input molecule.
To train the Mol-CycleGAN we use the following loss function:
and aim to solve
We use the adversarial loss introduced in LS-GAN [11]:
which ensures that the generator G (and F) generates samples from a distribution close to the distribution of Y (or X).
The cycle consistency loss:
reduces the space of possible mapping functions, such that for a molecule x from set X, the GAN cycle brings it back to a molecule similar to x, i.e. F(G(x)) is close to x (and analogously G(F(y)) is close to y).
Finally, to ensure that the generated (optimized) molecule is close to the starting one, we use the identity mapping loss [19]:
which further reduces the space of possible mapping functions and prevents the model from generating molecules that lay far away from the starting molecule in the latent space of JT-VAE.
In our experiments, we use the hyperparameters \(\lambda _1 = 0.3\) and \(\lambda _2 = 0.1\). Note that these parameters control the balance between improvement in the optimized property and similarity between the generated and the starting molecule.
2.1 Workflow
We conduct experiments to test if the proposed model is able to generate molecules that are close to the starting ones and possess increased octanol-water partition coefficient (logP) penalized by the synthetic accessibility (SA) score. We optimize penalized logP, while constraining the degree of deviation from the starting molecule. The similarity between molecules is measured with Tanimoto similarity on Morgan Fingerprints [14].
We use the ZINC-250K dataset used in similar studies [7, 10] which contains 250000 drug-like molecules extracted from the ZINC database [16]. The sets \(X_{\text {train}}\) and \(Y_{\text {train}}\) are random samples of size 80000 from ZINC-250K, where the compounds’ penalized logP values are below and above the median, respectively. \(X_{\text {test}}\) is a separate, non-overlapping dataset, consisting of 800 molecules with the lowest values of penalized logP in ZINC-250K.
All networks are trained using the Adam optimizer [8] with learning rate 0.0001, batch normalization [6] and leaky-ReLU with \(\alpha = 0.1\). The models are trained for 300 epochs. Generators are built of four fully connected residual layers, with 56 units. Discriminators are built of 7 dense layers of the following sizes: 48, 36, 28, 18, 12, 7, 1 units.
3 Results
We optimize the penalized logP under the constraint that the similarity between the original and the generated molecule is higher than a fixed threshold (denoted as \(\delta \)). This is a realistic scenario in drug discovery, where the development of new drugs usually starts with known molecules such as existing drugs [2].
We maximize the penalized logP coefficient and use the Tanimoto similarity with the Morgan fingerprint to define the threshold of similarity. We compare our results with previous similar studies [7, 18].
In our optimization procedure, each molecule is fed into the generator to obtain the ‘optimized’ molecule G(x). The pair (x, G(x)) defines an ’optimization path’ in the latent space of JT-VAE. To be able to make a comparison with the previous research [7] we start the procedure from the 800 molecules with the lowest values of penalized logP in ZINC-250K and then we decode molecules from 80 points along the path from x to G(x) in equal steps. From the resulting set of molecules we report the molecule with the highest penalized logP score that satisfies the similarity constraint. In the task of optimizing penalized logP of drug-like molecules, our method significantly outperforms the previous results in the mean improvement of the property (Table 1) and achieves a comparable mean similarity in the constrained scenario (for \(\delta > 0\)).
Molecules with highest improvement of logP are presented in Fig. 1 with the improvement given below the generated molecules.
Figure 2 shows starting and final molecules, together with all molecules generated along the optimization path and their values of penalized logP.
4 Conclusions
In this work, we introduce Mol-CycleGAN – a new model based on CycleGAN which can be used for the de novo generation of molecules. The advantage of the proposed model is the ability to learn transformation rules from the sets of compounds with desired and undesired values of the considered property. The model can generate molecules with desired properties, as shown on the example of penalized logP. The generated molecules are close to the starting ones and the degree of similarity can be controlled via a hyperparameter. In the task of constrained optimization of drug-like molecules our model significantly outperforms previous results.
The code used to produce the reported results can be found online at https://github.com/ardigen/mol-cycle-gan.
References
Bajorath, J.: Integration of virtual and high-throughput screening. Nat. Rev. Drug Discov. 1(11), 882–894 (2002). https://doi.org/10.1038/nrd941
Besnard, J., et al.: Automated design of ligands to polypharmacological profiles. Nature 492(7428), 215 (2012). https://doi.org/10.1038/nature11691
Bjerrum, E.J., Threlfall, R.: Molecular generation with recurrent neural networks (RNNs). arXiv preprint. arXiv:1705.04612 (2017)
Chen, H., Engkvist, O., Wang, Y., Olivecrona, M., Blaschke, T.: The rise of deep learning in drug discovery. Drug Discov. Today 23(6), 1241–1250 (2018). https://doi.org/10.1016/j.drudis.2018.01.039
Gupta, A., Müller, A.T., Huisman, B.J., Fuchs, J.A., Schneider, P., Schneider, G.: Generative recurrent networks for de novo drug design. Mol. Inform. 37(1–2), 1700111 (2018). https://doi.org/10.1002/minf.201700111
Ioffe, S., Szegedy, C.: Batch normalization: accelerating deep network training by reducing internal covariate shift. In: Proceedings of the 32nd International Conference on Machine Learning, vol. 37, pp. 448–456. ICML 2015, JMLR.org (2015). http://dl.acm.org/citation.cfm?id=3045118.3045167
Jin, W., Barzilay, R., Jaakkola, T.: Junction tree variational autoencoder for molecular graph generation. In: Dy, J., Krause, A. (eds.) Proceedings of the 35th International Conference on Machine Learning. Proceedings of Machine Learning Research, vol. 80, pp. 2323–2332. PMLR, Stockholmsmässan, Stockholm (10–15 July 2018)
Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization. arXiv preprint. arXiv:1412.6980 (2014)
Kingma, D.P., Welling, M.: Auto-encoding variational bayes. arXiv preprint. arXiv:1312.6114 (2013)
Kusner, M.J., Paige, B., Hernández-Lobato, J.M.: Grammar variational autoencoder. In: Proceedings of the 34th International Conference on Machine Learning, vol. 70, pp. 1945–1954. JMLR.org (2017)
Mao, X., Li, Q., Xie, H., Lau, R.Y., Wang, Z., Paul Smolley, S.: Least squares generative adversarial networks. In: 2017 IEEE International Conference on Computer Vision (ICCV), pp. 2794–2802 (2017). https://doi.org/10.1109/ICCV.2017.304
Rao, V.S., Srinivas, K.: Modern drug discovery process: an in silico approach. J. Bioinform. Seq. Anal. 2(5), 89–94 (2011)
Ratti, E., Trist, D.: The continuing evolution of the drug discovery process in the pharmaceutical industry. Farmaco 56(1–2), 13–19 (2001). https://doi.org/10.1016/S0014-827X(01)01019-9
Rogers, D., Hahn, M.: Extended-connectivity fingerprints. J. Chem. Inf. Model. 50(5), 742–754 (2010). https://doi.org/10.1021/ci100050t
Segler, M.H., Kogej, T., Tyrchan, C., Waller, M.P.: Generating focused molecule libraries for drug discovery with recurrent neural networks. ACS Cent. Sci. 4(1), 120–131 (2017). https://doi.org/10.1021/acscentsci.7b00512
Sterling, T., Irwin, J.J.: Zinc 15-ligand discovery for everyone. J. Chem. Inf. Model. 55(11), 2324–2337 (2015). https://doi.org/10.1021/acs.jcim.5b00559
Winter, R., Montanari, F., Noé, F., Clevert, D.A.: Learning continuous and data-driven molecular descriptors by translating equivalent chemical representations. Chem. Sci. 10(6), 1692–1701 (2019). https://doi.org/10.1039/C8SC04175J
You, J., Liu, B., Ying, Z., Pande, V., Leskovec, J.: Graph convolutional policy network for goal-directed molecular graph generation. In: Advances in Neural Information Processing Systems, pp. 6410–6421 (2018)
Zhu, J.Y., Park, T., Isola, P., Efros, A.A.: Unpaired image-to-image translation using cycle-consistent adversarial networks. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 2223–2232 (2017)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Open Access This chapter is licensed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license and indicate if changes were made.
The images or other third party material in this chapter are included in the chapter's Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the chapter's Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder.
Copyright information
© 2019 The Author(s)
About this paper
Cite this paper
Maziarka, Ł., Pocha, A., Kaczmarczyk, J., Rataj, K., Warchoł, M. (2019). Mol-CycleGAN - A Generative Model for Molecular Optimization. In: Tetko, I., Kůrková, V., Karpov, P., Theis, F. (eds) Artificial Neural Networks and Machine Learning – ICANN 2019: Workshop and Special Sessions. ICANN 2019. Lecture Notes in Computer Science(), vol 11731. Springer, Cham. https://doi.org/10.1007/978-3-030-30493-5_77
Download citation
DOI: https://doi.org/10.1007/978-3-030-30493-5_77
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-30492-8
Online ISBN: 978-3-030-30493-5
eBook Packages: Computer ScienceComputer Science (R0)