TAGMOL - Target Aware Gradient-Guided Molecule Generation
TAGMOL - Target Aware Gradient-Guided Molecule Generation
Vineeth Dorna * 1 2 D. Subhalingam * 1 Keshav Kolluru 1 Shreshth Tuli 1 3 Mrityunjay Singh 1 Saurabh Singal 1
N. M. Anoop Krishnan 4 Sayan Ranu 5
1
TAGM OL: Target-Aware Gradient-guided Molecule Generation
Preprint 2
TAGM OL: Target-Aware Gradient-guided Molecule Generation
Prior Guidance
Property
Predictor
Property
Predictor
Diffusion
Model
Loss
Denoising
Figure 1. Overview of TAGM OL. (a) Training a property-oriented guide using existing data. (b) Utilizing the trained guide and diffusion
model to steer the diffusion sampling process towards the optimal regions of the property of interest.
where N and C denotes the normal and categorical distri- 1
∥µ̃t XM M
− µ̃θ [XM M
(x) 2
Lt−1 = t , X0 t , Vt ], P, t ∥ + C
bution respectively while β1 , ..., βT represents the variance 2 β̃t
Preprint 3
TAGM OL: Target-Aware Gradient-guided Molecule Generation
dients ∇XMt
pϕ (y|Mt , P, t) to guide the diffusion sampling Once we get the final hidden states hLi,t , we predict the final
process towards the desirable properties encoded in y. property using an MLP layer as:
!
Fig 1 presents the pipeline of TAGM OL. TAGM OL con- f
X
L
ŷ = ϕy hi,t (8)
sists of two key components. The first component is Guide
i∈M
Training, introduced in § 3.1, where we present our SE(3) In-
variant GNN architecture and its modeling objectives, estab- While equivariance is essential for generative models to
lishing the foundation for our approach. The goal of this step maintain the consistency of the probability p(M0 |P) against
is to assess how well a molecule conforms to the desirable SE(3) transformations in protein-ligand complexes, it is also
properties Y. The second component, discussed in § 3.2, is vital that a scalar property predicted from guide remains
Guided Sampling. Here, we elaborate on how a trained guide independent of SE(3) transformations. Hence, the GNN in
model efficiently steers the diffusion sampling process to- guide is SE(3) invariant. This modeling approach introduces
wards specific regions of interest, emphasizing desired prop- a valuable inductive bias, which is particularly beneficial
erties. Our study goes further to demonstrate the concurrent in the generation of 3D objects, such as molecules (Köhler
optimization of multiple properties, as detailed in § 3.3. This et al., 2020; Satorras et al., 2021; Xu et al., 2022; Hooge-
multifaceted approach more holistically addresses the var- boom et al., 2022; Guan et al., 2023a;b).
ious aspects of molecular design, ultimately simplifying the
In our approach, diverging from the classifier-guided
efforts of chemists devote to the lead optimization process.
methodology (Dhariwal & Nichol, 2021) in diffusion pro-
3.1. Property Guide cesses where the classifier directly estimates probabilities,
We approach the input space as a 3D point cloud system we train our model to regress specific properties. The proba-
where we build a k-nearest neighbors (k-NN) graph G by bility modeling is articulated through a normal distribution
representing ligand and protein atoms as nodes, and each as:
atom is connected to its k-nearest neighbors. For properties pϕy (y|Mt , P) = N (y, ϕy (Mt , P, t), I) (9)
such as binding affinity, including protein atoms within the The training of these models is oriented toward minimiz-
graph is critical; however, the inclusion of protein atoms ing the Negative log-likelihood (N LL), which effectively
can be omitted for properties solely dependent on the ligand. simplifies the final objective to minimizing the Root Mean
We parameterize our guide model using an Invariant GNN Square Error (RMSE) between y and ϕy (Mt , P, t).
ϕy , which is trained to predict property y ∈ Y given a noisy
input Mt , protein P and time t. Later in the denoising T
X
phase, the gradients of ϕy are used to direct exploration N LL = −Ep(P,M0:T ) log(pϕy (y|Mt , P, t))
in regions of interest. Given the graph representation G at t=0
(10)
diffusion time step t, we define the GNN convolution layer T
X (y − ϕy (Mt , P, t))2
as follows: = Ep(P,M0:T )
2
di,j = D(∥xi,t − xj,t ∥ ) t=0
2
mi,j = ϕm l l
y (hi,t , hj,t , di,j , ai,j )
e Although we designed our guide for regression tasks, it is
ei,j = ϕy (mi,j )
X versatile enough to be adapted for classification tasks or
mi = ei,j mi,j other suitable applications.
j∈N (i)
(7)
3.2. Single Objective Guidance
hl+1
i,t = ϕhy (hli,t , mi )
( In this section, we first design the framework to obtain
ϕpy (viP ) if i ∈ P guidance from a single property. Subsequently, in § 3.3, we
h0i,t = M
ϕly ([vi,t , τ ]) if i ∈ M extend to multi-objective guidance.
Departing from the methodologies outlined in prior studies
where hli,t ∈ Rd represents the SE(3)-invariant hidden
(Guan et al., 2023a;b), our generative model distinctively
representation of protein and ligand atoms after l layers;
conditions on both the protein pocket and property to be
xi,t , xj,t ∈ R3 represents hidden representation of protein
guided. Remarkably, without retraining the diffusion model,
and ligand atoms. di,j represents the distance embedding
we guide the diffusion sampling process by shifting coor-
and ai,j represents the edge attributes in the graph.
dinates. This approach empowers the model to consistently
Invariance of hli,t stems from the fact that we take the
condition on optimal property values during the denoising
L2-distance between the atom representations. N (i) stands
phase. By steering the denoising process towards an optimal
for the set of neighbors for atom i. τ represents the time
property, it significantly enhances the model’s capability
embedding to make the model aware of noise at time step t.
to generate molecules with desired characteristics.
ϕm e h p l
y , ϕy , ϕy , ϕy and ϕy are Multi-Layer Perceptrons (MLP)
where as D is a distance encoder. As demonstrated in (Dhariwal & Nichol, 2021), efficiently
Preprint 4
TAGM OL: Target-Aware Gradient-guided Molecule Generation
sampling Mt−1 for each denoising transition can be ade- posterior distribution in Equation 12 as:
quately achieved by:
XM
t−1 ∼ N (µ̃θ (Mt , P, t) + δ, β̃I)
(15)
pθ,ϕy (Mt−1 |Mt , P, y) = Zpθ (Mt−1 |Mt , P) M
Vt−1 ∼ C (c˜θ (Mt , P, t))
· pϕy (y|Mt−1 , P, t − 1)
X
where δ= sy β̃∇XM log pϕy (y|Mt , P, t) (16)
(11) y∈Y
t
Preprint 5
TAGM OL: Target-Aware Gradient-guided Molecule Generation
Preprint 6
TAGM OL: Target-Aware Gradient-guided Molecule Generation
Table 1. Comparison of various properties between reference molecules and those generated by our model and other baselines. (↑) / (↓)
indicates whether a larger or smaller number is preferable. The first and second-place results are emphasized with bold and underlined
text, respectively. Refer to Appendix E.6 for evaluation on additional metrics suggested in (Harris et al., 2023).
Methods Vina Score (↓) Vina Min (↓) Vina Dock (↓) High Affinity (↑) QED (↑) SA (↑) Diversity (↑) Hit(↑)
Avg. Med. Avg. Med. Avg. Med. Avg. Med. Avg. Med. Avg. Med. Avg. Med. Rate %
Reference -6.36 -6.46 -6.71 -6.49 -7.45 -7.26 - - 0.48 0.47 0.73 0.74 - - 21
liGAN - - - - -6.33 -6.20 21.1% 11.1% 0.39 0.39 0.59 0.57 0.66 0.67 13.2
AR -5.75 -5.64 -6.18 -5.88 -6.75 -6.62 37.9% 31.0% 0.51 0.50 0.63 0.63 0.70 0.70 12.9
Pocket2Mol -5.14 -4.70 -6.42 -5.82 -7.15 -6.79 48.4% 51.0% 0.56 0.57 0.74 0.75 0.69 0.71 24.3
TargetDiff -5.47 -6.30 -6.64 -6.83 -7.80 -7.91 58.1% 59.1% 0.48 0.48 0.58 0.58 0.72 0.71 20.5
DecompDiff -4.85 -6.03 -6.76 -7.09 -8.48 -8.50 64.8% 78.6% 0.44 0.41 0.59 0.59 0.63 0.62 24.9
TAGM OL -7.02 -7.77 -7.95 -8.07 -8.59 -8.69 69.8% 76.4% 0.55 0.56 0.56 0.56 0.69 0.70 27.7
Density
0.3
Bond liGAN AR Pocket2 Target Decomp TAGM OL
Mol Diff Diff 0.2
C-C 0.601 0.609 0.496 0.369 0.371 0.384
0.1
C=C 0.665 0.620 0.561 0.505 0.539 0.501
C-N 0.634 0.474 0.416 0.363 0.352 0.365 0.0
C=N 0.749 0.635 0.629 0.550 0.592 0.559 0 2 4 6 8 10 12
C-O 0.656 0.492 0.454 0.421 0.373 0.422
Distance of all atom pairs (Å)
C=O 0.661 0.558 0.516 0.461 0.381 0.430
C:C 0.497 0.451 0.416 0.263 0.258 0.269 Figure 2. Comparison of distance distributions between all-atom
C:N 0.638 0.551 0.487 0.235 0.273 0.252 distances of reference molecules in the test set (Reference) and
distances in model-generated molecules. The Jensen-Shannon
divergence (JSD) between these two distributions is calculated and
reported.
4.5. Results
4.5.1. P ERFORMANCE OF TAGM OL able performance, surpassing diffusion-based models in
QED by 14.6% and 22.2% compared to TargetDiff and De-
TAGM OL outperforms all baselines, including the refer-
compDiff, respectively, while maintaining similar diversity
ence molecules, in binding-related metrics and hit rate (see
levels. Although there’s a slight decrease in SA, the metric
Table 1). This achievement is especially noteworthy when
remains within a reasonable range, indicating satisfactory
compared to the state-of-the-art DecompDiff model, which
synthesizability. A detailed discussion of the challenges
relies on external computation for informed priors in the de-
in optimizing SA is provided in Appendix E.4. Moreover,
noising process. The success of our model, attained without
Appendix E.5 presents compelling evidence of statistically
requiring extra data, underscores the importance of effec-
significant changes in guided properties upon the incorpo-
tively learning useful signals from the existing training set
ration of guidance. Furthermore, as shown in Table 2, our
and skillfully guiding the diffusion denoising phase.
method closely maintains the bond distribution in align-
A key metric where TAGM OL excels is the Vina Score, ment with the backbone diffusion model, namely TargetDiff,
achieving a 22% improvement over the state of the art surpassing the non-diffusion baselines. This outcome em-
(AR). This highlights its proficiency not only in generating phasizes the efficient synergy between our guide model and
molecules that bind effectively with proteins but also in the diffusion model, markedly improving our ability to gen-
producing high-affinity poses. This is further substantiated erate molecules with targeted properties while preserving
by the fact that on average an impressive 69.8% of the molecular conformation.
molecules generated by our model exhibit superior binding
TAGM OL, which focuses on optimizing multiple properties,
affinity compared to reference molecules, surpassing all
achieves a state-of-the-art hit rate, surpassing the backbone
other baselines.
model—TargetDiff—by a considerable margin. This is ev-
In terms of molecular properties, TAGM OL shows remark- idenced by the higher fraction of molecules generated by
Preprint 7
TAGM OL: Target-Aware Gradient-guided Molecule Generation
Table 3. Ablation analysis assessing the properties of generated molecules under different property guidance scenarios. The first and
second-place results are emphasized with bold and underlined text, respectively.
Methods Vina Score (↓) Vina Min (↓) Vina Dock (↓) QED (↑) SA (↑) Hit (↑)
Avg. Med. Avg. Med. Avg. Med. Avg. Med. Avg. Med. Rate
backbone -5.47 -6.30 -6.64 -6.83 -7.80 -7.91 0.48 0.48 0.58 0.58 20.5
backbone + BA Opt -7.35 -8.18 -8.38 -8.46 -9.04 -9.04 0.49 0.50 0.53 0.53 22.6
backbone + QED Opt -5.48 -6.46 -6.77 -6.93 -7.93 -8.06 0.56 0.57 0.58 0.58 24.5
backbone + SA Opt -5.22 -6.03 -6.40 -6.57 -7.53 -7.73 0.47 0.48 0.61 0.60 19.4
TAGM OL -7.02 -7.77 -7.95 -8.07 -8.59 -8.69 0.55 0.56 0.56 0.56 27.7
Preprint 8
TAGM OL: Target-Aware Gradient-guided Molecule Generation
directions worth investigating in the future. First, improved and Hopkins, A. L. Quantifying the chemical beauty of
representation of the system using additional inductive drugs. Nature chemistry, 4(2):90–98, 2012.
bias in terms of symmetry and coarse-graining could
furnish more accurate predictions of different properties Bjerrum, E. J. and Threlfall, R. Molecular generation with
that are structure dependent. Second, model compression recurrent neural networks (rnns). ArXiv, abs/1705.04612,
and parameter pruning could speed-up inference without 2017. URL https://api.semanticscholar.
compromising performance. Third, integrating constraints org/CorpusID:8280382.
on discrete attributes into gradient-based optimization may Chenthamarakshan, V., Das, P., Hoffman, S., Strobelt, H.,
enable the modeling of complementary properties, such as Padhi, I., Lim, K. W., Hoover, B., Manica, M., Born, J.,
enforcing hydrogen bonds between specific residues within Laino, T., et al. Cogmol: Target-specific and selective
the pocket, a capability our current algorithm lacks. Fourth, drug design for covid-19 using deep generative models.
exploring reduced noise sampling strategies or guidance Advances in Neural Information Processing Systems, 33:
mechanisms for generating more precise conformations 4320–4332, 2020.
could lead to reduced strain energy and the production of
stable molecules. Finally, it would be worth investigating Dhariwal, P. and Nichol, A. Diffusion models beat gans
physics-based techniques, such as molecular simulations, on image synthesis. Advances in neural information
to allow generalization to larger family of systems. processing systems, 34:8780–8794, 2021.
Eberhardt, J., Santos-Martins, D., Tillack, A. F., and Forli, S.
6. Broader Impact Autodock vina 1.2. 0: New docking methods, expanded
force field, and python bindings. Journal of chemical
Drug discovery is an area that can have broad impact on
information and modeling, 61(8):3891–3898, 2021.
humanity. Accelerating the discovery of novel drugs can en-
sure more inclusive healthcare worldwide and can have im- Eckmann, P., Sun, K., Zhao, B., Feng, M., Gilson, M. K.,
pact especially on reducing the disparity within and across and Yu, R. Limo: Latent inceptionism for targeted
countries. In the development of our methodology, we have molecule generation. arXiv preprint arXiv:2206.09010,
taken care to ensure its ethical application, fully acknowl- 2022.
edging that there exists inadvertent potential for misuse.
Specifically, there is a risk that our optimization algorithms Ertl, P. and Schuffenhauer, A. Estimation of synthetic acces-
could be repurposed for creating drugs with detrimental side sibility score of drug-like molecules based on molecular
effects. This underscores the importance of exercising cau- complexity and fragment contributions. Journal of chem-
tion and responsibility in the deployment of such algorithms informatics, 1:1–11, 2009.
in practical scenarios to prevent any unintended negative
Francoeur, P. G., Masuda, T., Sunseri, J., Jia, A., Iovanisci,
consequences from their use.
R. B., Snyder, I., and Koes, D. R. Three-dimensional
convolutional neural networks and a cross-docked data
References set for structure-based drug design. Journal of chemical
information and modeling, 60(9):4200–4215, 2020.
Alonso, H., Bliznyuk, A. A., and Gready, J. E. Com-
bining docking and molecular dynamic simulations Ghorbani, M., Gendelev, L., Beroza, P., and Keiser, M. Au-
in drug design. Medicinal Research Reviews, 26, toregressive fragment-based diffusion for pocket-aware
2006. URL https://api.semanticscholar. ligand design. In NeurIPS 2023 Generative AI and Biol-
org/CorpusID:16337213. ogy (GenBio) Workshop, 2023.
Anderson, A. C. The process of structure-based drug design. Gómez-Bombarelli, R., Wei, J. N., Duvenaud, D.,
Chemistry & biology, 10(9):787–797, 2003. Hernández-Lobato, J. M., Sánchez-Lengeling, B., She-
Bajusz, D., Rácz, A., and Héberger, K. Why is tanimoto berla, D., Aguilera-Iparraguirre, J., Hirzel, T. D., Adams,
index an appropriate choice for fingerprint-based similar- R. P., and Aspuru-Guzik, A. Automatic chemical de-
ity calculations? Journal of cheminformatics, 7(1):1–13, sign using a data-driven continuous representation of
2015. molecules. ACS central science, 4(2):268–276, 2018.
Bao, F., Zhao, M., Hao, Z., Li, P., Li, C., and Zhu, J. Equiv- Guan, J., Qian, W. W., Peng, X., Su, Y., Peng, J., and Ma,
ariant energy-guided sde for inverse molecular design. In J. 3d equivariant diffusion for target-aware molecule
The eleventh international conference on learning repre- generation and affinity prediction. In The Eleventh In-
sentations, 2022. ternational Conference on Learning Representations,
2023a. URL https://openreview.net/forum?
Bickerton, G. R., Paolini, G. V., Besnard, J., Muresan, S., id=kJqXEPXMsE0.
Preprint 9
TAGM OL: Target-Aware Gradient-guided Molecule Generation
Guan, J., Zhou, X., Yang, Y., Bao, Y., Peng, J., Ma, J., Liu, Liu, Q., Allamanis, M., Brockschmidt, M., and Gaunt, A. L.
Q., Wang, L., and Gu, Q. Decompdiff: Diffusion models Constrained graph variational autoencoders for molecule
with decomposed priors for structure-based drug design. design. In Neural Information Processing Systems,
2023b. 2018. URL https://api.semanticscholar.
org/CorpusID:43924638.
Harris, C., Didi, K., Jamasb, A., Joshi, C., Mathis, S., Lio,
P., and Blundell, T. Posecheck: Generative models for Liu, Z., Zhang, W., Xia, Y., Wu, L., Xie, S., Qin, T., Zhang,
3d structure-based drug design produce unrealistic poses. M., and Liu, T.-Y. MolXPT: Wrapping molecules with
In NeurIPS 2023 Generative AI and Biology (GenBio) text for generative pre-training. In Proceedings of the 61st
Workshop, 2023. Annual Meeting of the Association for Computational
Linguistics (Volume 2: Short Papers), Toronto, Canada,
Ho, J., Jain, A., and Abbeel, P. Denoising diffusion proba- July 2023. Association for Computational Linguistics.
bilistic models. Advances in Neural Information Process-
Long, S., Zhou, Y., Dai, X., and Zhou, H. Zero-shot 3d drug
ing Systems, 33:6840–6851, 2020.
design by sketching and generating. Advances in Neural
Hoogeboom, E., Satorras, V. G., Vignac, C., and Welling, M. Information Processing Systems, 35:23894–23907, 2022.
Equivariant diffusion for molecule generation in 3d. In Luo, S., Guan, J., Ma, J., and Peng, J. A 3d generative model
International conference on machine learning, pp. 8867– for structure-based drug design. Advances in Neural
8887. PMLR, 2022. Information Processing Systems, 34:6229–6239, 2021.
Huang, Z., Yang, L., Zhou, X., Zhang, Z., Zhang, W., Zheng, Olivecrona, M., Blaschke, T., Engkvist, O., and Chen, H.
X., Chen, J., Wang, Y., CUI, B., and Yang, W. Protein- Molecular de-novo design through deep reinforcement
ligand interaction prior for binding-aware 3d molecule learning. Journal of cheminformatics, 9(1):1–14, 2017.
diffusion models. In The Twelfth International Confer-
Peng, X., Luo, S., Guan, J., Xie, Q., Peng, J., and Ma, J.
ence on Learning Representations, 2024. URL https:
Pocket2mol: Efficient molecular sampling based on 3d
//openreview.net/forum?id=qH9nrMNTIW.
protein pockets. In International Conference on Machine
Jeon, W. and Kim, D. Autonomous molecule generation Learning, pp. 17644–17655. PMLR, 2022.
using reinforcement learning and docking to develop po- Peng, X., Guan, J., Liu, Q., and Ma, J. Moldiff: Addressing
tential novel inhibitors. Scientific reports, 10(1):22104, the atom-bond inconsistency problem in 3d molecule
2020. diffusion generation. arXiv preprint arXiv:2305.07508,
2023.
Katigbak, J., Li, H., Rooklin, D., and Zhang, Y. Alphaspace
2.0: representing concave biomolecular surfaces using β- Ragoza, M., Masuda, T., and Koes, D. R. Learning a
clusters. Journal of chemical information and modeling, continuous representation of 3d molecular structures
60(3):1494–1508, 2020. with deep generative models. ArXiv, abs/2010.08687,
2020. URL https://api.semanticscholar.
Kingma, D. P. and Welling, M. Auto-encoding variational org/CorpusID:224713572.
bayes. arXiv preprint arXiv:1312.6114, 2013.
Ragoza, M., Masuda, T., and Koes, D. R. Generating 3d
Köhler, J., Klein, L., and Noé, F. Equivariant flows: exact molecules conditional on receptor binding sites with deep
likelihood generative learning for symmetric densities. generative models. Chemical science, 13(9):2701–2713,
In International Conference on Machine Learning, pp. 2022.
5361–5370. PMLR, 2020. Satorras, V. G., Hoogeboom, E., Fuchs, F. B., Posner, I., and
Welling, M. E (n) equivariant normalizing flows. arXiv
Landrum, G. Rdkit documentation. Release, 1(1-79):4,
preprint arXiv:2105.09016, 2021.
2013.
Skalic, M., Jiménez, J., Sabbadin, D., and De Fabritiis,
Lee, S., Jo, J., and Hwang, S. J. Exploring chemical space G. Shape-based generative modeling for de novo drug
with score-based out-of-distribution generation. Proceed- design. Journal of chemical information and modeling,
ings of the 40th International Conference on Machine 59(3):1205–1214, 2019.
Learning, 2023.
Sohl-Dickstein, J., Weiss, E., Maheswaranathan, N., and
Liu, M., Luo, Y., Uchino, K., Maruhashi, K., and Ji, S. Ganguli, S. Deep unsupervised learning using nonequi-
Generating 3d molecules for target protein binding. In librium thermodynamics. In International conference on
International Conference on Machine Learning, 2022. machine learning, pp. 2256–2265. PMLR, 2015.
Preprint 10
TAGM OL: Target-Aware Gradient-guided Molecule Generation
Stanton, S., Maddox, W., Gruver, N., Maffettone, P., De- Zhou, X., Cheng, X., Yang, Y., Bao, Y., Wang, L., and
laney, E., Greenside, P., and Wilson, A. G. Accelerating Gu, Q. Decompopt: Controllable and decomposed diffu-
bayesian optimization for biological sequence design with sion models for structure-based molecular optimization.
denoising autoencoders. In International Conference on In The Twelfth International Conference on Learning
Machine Learning, pp. 20459–20478. PMLR, 2022. Representations, 2024. URL https://openreview.
net/forum?id=Y3BbxvAQS9.
Tanimoto, T. T. Elementary mathematical theory of classifi-
cation and prediction. 1958. Zhou, Z., Kearnes, S. M., Li, L., Zare, R. N., and Riley, P. F.
Optimization of molecules via deep reinforcement learn-
Trott, O. and Olson, A. J. Autodock vina: improving the ing. Scientific Reports, 9, 2018. URL https://api.
speed and accuracy of docking with a new scoring func- semanticscholar.org/CorpusID:53040150.
tion, efficient optimization, and multithreading. Journal
Zhou, Z., Kearnes, S., Li, L., Zare, R. N., and Riley, P. Op-
of computational chemistry, 31(2):455–461, 2010.
timization of molecules via deep reinforcement learning.
Scientific reports, 9(1):10752, 2019.
Xu, M., Yu, L., Song, Y., Shi, C., Ermon, S., and Tang, J.
Geodiff: A geometric diffusion model for molecular con-
formation generation. arXiv preprint arXiv:2203.02923,
2022.
Yang, K., Swanson, K., Jin, W., Coley, C. W., Eiden, P.,
Gao, H., Guzman-Perez, A., Hopper, T., Kelley, B. P.,
Mathea, M., Palmer, A., Settels, V., Jaakkola, T., Jensen,
K. F., and Barzilay, R. Analyzing learned molecu-
lar representations for property prediction. Journal of
Chemical Information and Modeling, 59:3370 – 3388,
2019. URL https://api.semanticscholar.
org/CorpusID:198986021.
Yang, Y., Ouyang, S., Dang, M., Zheng, M., Li, L., and
Zhou, H. Knowledge guided geometric editing for unsu-
pervised drug design. 2021.
You, J., Liu, B., Ying, Z., Pande, V., and Leskovec, J. Graph
convolutional policy network for goal-directed molecu-
lar graph generation. Advances in neural information
processing systems, 31, 2018.
Zhang, Z., Min, Y., Zheng, S., and Liu, Q. Molecule gen-
eration for target protein binding with structural motifs.
In The Eleventh International Conference on Learning
Representations, 2023. URL https://openreview.
net/forum?id=Rq13idF0F73.
Zhou, G., Gao, Z., Ding, Q., Zheng, H., Xu, H., Wei, Z.,
Zhang, L., and Ke, G. Uni-mol: A universal 3d molecu-
lar representation learning framework. In The Eleventh
International Conference on Learning Representations,
2023. URL https://openreview.net/forum?
id=6K2RM6wVqKu.
Preprint 11
TAGM OL: Target-Aware Gradient-guided Molecule Generation
A. Notation Summary
Table 4. Notations
B. Related Work
Neural Drug Design Neural models have significantly changed the landscape for drug design that was previously
dominated by computational techniques such as MD simulations and Docking (Alonso et al., 2006). It has gone through
multiple paradigm shifts in the recent past - ranging from using only 1D representations (SMILES) (Bjerrum & Threlfall,
2017; Gómez-Bombarelli et al., 2018) to 2D representations (molecular graphs) (Liu et al., 2018; Zhou et al., 2018) to 3D
representations (Skalic et al., 2019; Ragoza et al., 2020) (molecular coordinates). Recent works have also showcased the
importance of using target-aware models for practical drug design. Advancements in target-aware drug design, particularly
in the realm of text and graph-based generative methods (Eckmann et al., 2022; Stanton et al., 2022; Chenthamarakshan
et al., 2020; Lee et al., 2023), have made significant strides. However, these methods often generate ligands without
considering the 3D structure of the target protein. TargetDiff (Guan et al., 2023a) and DecompDiff (Guan et al., 2023b)
are two diffusion-based models that consider the 3D structure of the target protein. DecompDiff extends TargetDiff by
decomposing the task into generating the arms and scaffold while explicitly considering bonds between the atoms during the
diffusion process which is ignored by TargetDiff. Fragment-based generation methods such as (Ghorbani et al., 2023) allow
for controlled generation, ensuring that only certain types of molecular fragments are present in the generated ligand. But
they still fall behind the performance of diffusion-based approaches in terms of their binding affinity.
Molecular Property Prediction Predicting the properties of molecules through physio-chemical experiments is an
expensive and time-consuming solution, that is unsuitable for getting intermediate feedback on AI-designed molecules.
Using neural models has allowed high-quality prediction of various molecular properties in an automated fashion. Neural
methods use different molecular representations to predict molecules, including SMILES representation with pre-trained
transformers (Liu et al., 2023) or molecular structures (Yang et al., 2019; Zhou et al., 2023) for predicting the properties of
interest.
Preprint 12
TAGM OL: Target-Aware Gradient-guided Molecule Generation
Molecular Property Optimization Prior methods for property optimization have focused on using Reinforcement learning
to generate molecules with desired properties (You et al., 2018; Zhou et al., 2019; Zhavoronkov et al., 2019; Jeon & Kim,
2020; Olivecrona et al., 2017). However, RL methods are often computationally expensive and challenging to optimize
due to the vast search space. LIMO (Eckmann et al., 2022) uses Variational Auto Encoders (VAEs) and gradient gudiance
on MLP-based property predictors but does not consider the 3D structure of the target protein or the generated ligand.
(Lee et al., 2023) introduced a novel graph-based diffusion model for out-of-distribution (OOD) generation, enhancing
explorability through OOD-controlled reverse-time diffusion and property-guided sampling, albeit focusing solely on 2D
molecular representations, unlike the target-aware 3D structure generation with multi-objective guidance featured in our
work.
6000 10000
12500
8000
10000
Frequency
4000
6000 7500
4000 5000
2000
2000 2500
0 0 0
0.0 0.2 0.4 0.6 0.8 0.2 0.4 0.6 0.8 1.0 17.5 15.0 12.5 10.0 7.5 5.0 2.5 0.0
QED SA Vina Score
Figure 4. Distribution of QED, SA, and Vina Score properties in the training split, with the mean values represented by dotted lines
D.2. Evaluation
The performance of each guide model was evaluated for their respective property predictions using the same test dataset. As
outlined in § 4.1, we employed the same split to evaluate our binding affinity guide model. While the train-test split ensures
the uniqueness of the protein-ligand complexes between the sets, 34 ligands are common to both, leading to potential train-
test leakage. Consequently, for QED and SA guides that depend solely on ligands, we have excluded these 34 overlapping
ligands from the test split (No overlaps). Table 5 presents the performance evaluation of each guide model, indicating high
R2 values on the original test dataset. After removing the overlapping ligands from the test set, the performance metrics
remained consistent, confirming the model’s effective generalization to unseen ligands. While we report the results on fair
test split, it’s crucial to note that, diverging from the traditional machine learning training and testing process, we train the
guide to direct the generated latent space of the diffusion model.
Preprint 13
TAGM OL: Target-Aware Gradient-guided Molecule Generation
Table 5. Evaluation of the guide model on the original test set with unique protein-ligand complexes, and on a modified test set (No
Overlaps) where 32 overlapping ligands have been removed.
E. Additonal Experiments
E.1. Search for optimal Guide Strength
To determine the optimal Guide Strength (s), a grid search was conducted over the set of values: {0, 1, 2, 5, 10, 20, 50}. Here,
s = 0 represents the scenario without guidance, included for comparative purposes to gauge guide-induced enhancements.
The generation process was limited to 50 molecules for a limited set of 4 targets, to accommodate computational constraints.
Initially, for each property, we find the optimal guide strength sopt
y over the grid. Once the optimal values for single objective
are tuned, for multi-objective guidance, we obtained optimal s values by recalibrating the sopt
y with wy , as detailed in §-4.4.
Table 6 presents the results of our tuning across different guide strengths (s) mentioned in the grid. It is evident that non-zero
s values yield improved property metrics, underscoring the efficacy of guidance. For Binding Affinity (BA) guidance,
optimal s values emerge as 1 and 2, based on average and median Vina Scores, respectively. Considering both Vina Min
and Vina Dock values, s = 2 is selected for superior outcomes. In the case of QED guidance, an s value of 20 distinctly
outperforms others in terms of both average and median QED scores. With SA guidance, although s = 20 and s = 50
appear to be optimal, the validity(Guan et al., 2023b) of generated molecules takes a low value of 67% and 46%, respectively.
Consequently, s = 5 is chosen, providing the second-best SA values, with 84% of molecules being valid and only a slight
decrease in average SA value by 0.02. In summary, the selected Guide Strengths are: sopt opt opt
ba = 2, sqed = 20, and ssa = 5.
Table 6. Search for optimal Guide Strength (s), conducted individually for each Guide Model (single-objective setting). To minimize
computational demands, the generation is limited to 50 samples across 4 targets. Only the metrics of the property being guided are
included. Symbols (↑) / (↓) denote preference for higher or lower values, respectively. Top results are highlighted in bold for first place
and underlined for second. Italicized figures indicate instances where over 25% of molecules generated under the specified guide strength
were not valid.
Preprint 14
TAGM OL: Target-Aware Gradient-guided Molecule Generation
properties. However, given the multifaceted nature of drug discovery, where the goal is to produce molecules with high
binding affinity and desired properties like QED and SA, our approach significantly enhances hit rate and most properties
over no guidance. This underscores the effectiveness of guidance in navigating regions that are crucial in discovery.
Table 7. Extensive ablation analysis assessing the properties of generated molecules under different property guidance scenarios. The first
and second-place results are emphasized with bold and underlined text, respectively.
wba wqed wsa Vina Score (↓) Vina Min (↓) Vina Dock (↓) QED (↑) SA (↑) Hit (↑)
Avg. Med. Avg. Med. Avg. Med. Avg. Med. Avg. Med. Rate
0 0 0 -5.47 -6.30 -6.64 -6.83 -7.80 -7.91 0.48 0.48 0.58 0.58 20.5
1 0 0 -7.35 -8.18 -8.38 -8.46 -9.04 -9.04 0.49 0.5 0.53 0.53 22.6
0 1 0 -5.48 -6.46 -6.77 -6.93 -7.93 -8.06 0.56 0.57 0.58 0.58 24.5
0 0 1 -5.22 -6.03 -6.4 -6.57 -7.53 -7.73 0.47 0.48 0.61 0.6 19.4
0.5 0.5 0 -7.11 -7.96 -8.13 -8.21 -8.82 -8.85 0.56 0.57 0.55 0.55 26.1
0.5 0 0.5 -7.2 -7.95 -8.16 -8.26 -8.83 -8.84 0.5 0.51 0.55 0.55 24.7
0 0.5 0.5 -5.43 -6.34 -6.63 -6.86 -7.85 -7.97 0.55 0.56 0.59 0.59 24.1
0.34 0.33 0.33 -7.02 -7.77 -7.95 -8.07 -8.59 -8.69 0.55 0.56 0.56 0.56 27.7
Density
Density
2.0
1.00
0.06
1.5
0.75
0.04
0.50 1.0
0.02 0.5
0.25
Figure 5. Distribution of molecular properties in molecules generated by the backbone model without any guidance (No Opt), when
guided by two properties while excluding the one indicated on the x-axis (Excluded Property on x-axis), and when guided by all three
properties (All Properties). (↓) denote properties where lower values are preferred, while (↑) indicate properties where higher values are
desirable.
Figure 5 presents the distribution of molecular properties for 10,000 generated molecules (100 per target, across 100 targets).
These molecules are generated using three methods: (i) the backbone model without any guidance; (ii) guidance using only
two properties, leaving out the third one indicated on the x-axis; and (iii) guidance using all three properties. Optimizing all
the properties helps improve the Vina Score and QED compared to the ‘No Opt’ baseline. Although it doesn’t help improve
the SA, it is still helpful, as excluding the SA property in optimization worsens the SA scores. Thus, it is useful to provide
guidance for all three properties.
Preprint 15
TAGM OL: Target-Aware Gradient-guided Molecule Generation
Table 8. Distribution of ring sizes in reference and generated molecules, expressed as percentages
Preprint 16
TAGM OL: Target-Aware Gradient-guided Molecule Generation
also observed in TAGMol due to minor errors in atom placement resulting from coordinate guidance. We hypothesize that
integrating strain energy guidance during the generation phase could yield more stable molecules. We plan to investigate
this in our future work.
Table 9. Evaluation of poses of generated molecules on biophysical benchmarks, PoseCheck. (↑) / (↓) indicates whether a larger or smaller
value is preferable. The first-place results are emphasized in bold.
Table 10. Evaluation of poses of generated molecules on biophysical benchmarks, PoseCheck, after redocking. (↑) / (↓) indicates whether
a larger or smaller value is preferable. The first-place results are emphasized in bold.
F. Time Complexity
When executed on an NVIDIA A100 GPU, we observed a notable processing time trend for inference. Guidance for
properties like QED and SA led to significant improvements (refer to Table 7) with minimal time increases of 1.06x and
1
Since these are contemporary works that will appear in ICLR 24, a comprehensive comparison could not be conducted.
Preprint 17
TAGM OL: Target-Aware Gradient-guided Molecule Generation
Table 11. Comparison of various properties between reference molecules and those generated by our model and other new baselines. (↑) /
(↓) indicates whether a larger or smaller number is preferable. The first and second-place results are emphasized with bold and underlined
text, respectively.
Methods Vina Score (↓) Vina Min (↓) Vina Dock (↓) High Affinity (↑) QED (↑) SA (↑) Diversity (↑)
Avg. Med. Avg. Med. Avg. Med. Avg. Med. Avg. Med. Avg. Med. Avg. Med.
Reference -6.36 -6.46 -6.71 -6.49 -7.45 -7.26 - - 0.48 0.47 0.73 0.74 - -
TargetDiff -5.47 -6.30 -6.64 -6.83 -7.80 -7.91 58.1% 59.1% 0.48 0.48 0.58 0.58 0.72 0.71
DecompDiff -4.85 -6.03 -6.76 -7.09 -8.48 -8.50 64.8% 78.6% 0.44 0.41 0.59 0.59 0.63 0.62
IPDiff -6.42 -7.01 -7.45 -7.48 -8.57 -8.51 69.5% 75.5% 0.52 0.52 0.59 0.59 0.74 0.73
DecompOPT -5.87 -6.81 -7.35 -7.78 -8.98 -9.01 73.5% 93.3% 0.48 0.45 0.65 0.65 0.60 0.60
TAGM OL -7.02 -7.77 -7.95 -8.07 -8.59 -8.69 69.8% 76.4% 0.55 0.56 0.56 0.56 0.68 0.69
1.07x, respectively, compared to the backbone. However, guiding for Binding Affinity (BA) led to a more substantial time
increase due to the inclusion of protein atoms, demanding extra computational effort compared to QED or SA guidance,
which involves only ligand atoms. Ultimately, TAGM OL recorded its longest processing time of 1.87x when applying
guidance for all properties. It is important to note that TAGM OL is still comparable to 1.72x of DecompDiff (Guan et al.,
2023a) sampling time.
Preprint 18