Disentangled autoencoding equivariant diffusion model for controlled generation of 3D molecules
Disentangled autoencoding equivariant diffusion model for controlled generation of 3D molecules
References
- Jin, W., Barzilay, R. & Jaakkola, T. Junction tree variational autoencoder for molecular graph generation. In International Conference on Machine Learning 2323–2332 (PMLR, 2018).
- Zang, C. & Wang, F. Moflow: an invertible flow model for generating molecular graphs. In Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, 617–626 (2020).
- Jin, W., Wohlwend, J., Barzilay, R. & Jaakkola, T. Iterative refinement graph neural network for antibody sequence-structure co-design. In International Conference on Learning Representations (ICLR, 2022).
- Luo, S. et al. Antigen-specific antibody design and optimization with diffusion-based generative models for protein structures. Adv. Neural Inf. Process. Syst. 35, 9754–9767 (2022).
Google Scholar
- Li, T., Guo, H., Grazioli, F., Gerstein, M. & Min, M. R. Disentangled Wasserstein autoencoder for T-cell receptor engineering. Adv. Neural Inf. Process. Syst. 36, 73604–73632 (2023).
- Fuchs, F., Worrall, D., Fischer, V. & Welling, M. Se (3)-transformers: 3D roto-translation equivariant attention networks. Adv. Neural Inf. Process. Syst. 33, 1970–1981 (2020).
Google Scholar
- Finzi, M., Stanton, S., Izmailov, P. & Wilson, A. G. Generalizing convolutional neural networks for equivariance to lie groups on arbitrary continuous data. In International Conference on Machine Learning, 3165–3176 (PMLR, 2020).
- Doerr, S. et al. Torchmd: a deep learning framework for molecular simulations. J. Chem. Theory Comput. 17, 2355–2363 (2020).
- Wu, F. & Li, S. Z. Diffmd: A geometric diffusion model for molecular dynamics simulations. In Proceedings of the AAAI Conference on Artificial Intelligence, Vol. 37, 5321–5329 (AAAI Press, 2023).
- Musaelian, A. et al. Learning local equivariant representations for large-scale atomistic dynamics. Nat. Commun. 14, 579 (2023).
Google Scholar
- Anderson, A. C. The process of structure-based drug design. Chem. Biol. 10, 787–797 (2003).
Google Scholar
- Lim, J. et al. Predicting drug–target interaction using a novel graph neural network with 3D structure-embedded graph representation. J. Chem. Inf. Model. 59, 3981–3988 (2019).
Google Scholar
- Anand, N. & Achim, T. Protein structure and sequence generation with equivariant denoising diffusion probabilistic models. arXiv preprint https://doi.org/10.48550/arXiv.2205.15019 (2022).
- Gebauer, N., Gastegger, M. & Schütt, K. Symmetry-adapted generation of 3d point sets for the targeted discovery of molecules. Advances in Neural Information Processing Systems 32 https://doi.org/10.48550/arXiv.1906.00957 (2019).
- Hoogeboom, E., Satorras, V. G., Vignac, C. & Welling, M. Equivariant diffusion for molecule generation in 3d. In International Conference on Machine Learning, 8867–8887 (PMLR, 2022).
- Xu, M. et al. Geodiff: a geometric diffusion model for molecular conformation generation. In International Conference on Learning Representations (ICLR, 2022).
- Xu, M., Powers, A., Dror, R., Ermon, S. & Leskovec, J. Geometric latent diffusion models for 3d molecule generation. In International Conference on Machine Learning, 38592–38610 (PMLR, 2023).
- Schneuing, A. et al. Structure-based drug design with equivariant diffusion models. Nat. Comput. Sci. 4, 899–909 (2024).
- Torge, J., Harris, C., Mathis, S. V. & Lio, P. Diffhopp: a graph diffusion model for novel drug design via scaffold hopping. arXiv preprint https://doi.org/10.48550/arXiv.2308.07416 (2023).
- Igashov, I. et al. Equivariant 3d-conditional diffusion model for molecular linker design. Nat. Mach. Intell. 6, 417–427 (2024).
- Ho, J., Jain, A. & Abbeel, P. Denoising diffusion probabilistic models. Adv. Neural Inf. Process. Syst. 33, 6840–6851 (2020).
Google Scholar
- Song, J., Meng, C. & Ermon, S. Denoising diffusion implicit models. In International Conference on Learning Representations (ICLR, 2021).
- Huang, H., Sun, L., Du, B. & Lv, W. Learning joint 2D & 3D diffusion models for complete molecule generation. arXiv preprint https://doi.org/10.48550/arXiv.2305.12347 (2023).
- Vignac, C., Osman, N., Toni, L. & Frossard, P. Midi: Mixed graph and 3D denoising diffusion for molecule generation. In Joint European Conference on Machine Learning and Knowledge Discovery in Databases, 560–576 (2023).
- Peng, X., Guan, J., Liu, Q. & Ma, J. Moldiff: Addressing the atom-bond inconsistency problem in 3d molecule diffusion generation. In International Conference on Machine Learning, 27611–27629 (PMLR, 2023).
- Gómez-Bombarelli, R. et al. Automatic chemical design using a data-driven continuous representation of molecules. ACS Cent. Sci. 4, 268–276 (2018).
Google Scholar
- Jin, W., Yang, K., Barzilay, R. & Jaakkola, T. Learning multimodal graph-to-graph translation for molecular optimization. In International Conference on Learning Representations (ICLR, 2018).
- Jin, W., Barzilay, R. & Jaakkola, T. Hierarchical generation of molecular graphs using structural motifs. In International Conference on Machine Learning, 4839–4848 (PMLR, 2020).
- Chen, Z., Min, M. R., Parthasarathy, S. & Ning, X. A deep generative model for molecule optimization via one fragment modification. Nat. Mach. Intell. 3, 1040–1049 (2021).
Google Scholar
- Du, Y. et al. Chemspace: Interpretable and interactive chemical space exploration. Transactions on Machine Learning Research (2022).
- Wang, Z. et al. Retrieval-based controllable molecule generation. In International Conference on Learning Representations (ICLR, 2023).
- Liu, S. et al. Multi-modal molecule structure–text model for text-based retrieval and editing. Nat. Mach. Intell. 5, 1447–1457 (2023).
Google Scholar
- Morehead, A. & Cheng, J. Geometry-complete diffusion for 3d molecule generation and optimization. Commun. Chem. 7, 150 (2024).
Google Scholar
- Cremer, J., Le, T., Noé, F., Clevert, D.-A. & Schütt, K. T. Pilot: equivariant diffusion for pocket-conditioned de novo ligand generation with multi-objective guidance via importance sampling. Chem. Sci. 15, 14954–14967 (2024).
Google Scholar
- Bao, F. et al. Equivariant energy-guided sde for inverse molecular design. In International Conference on Learning Representations (ICLR, 2023).
- Rombach, R., Blattmann, A., Lorenz, D., Esser, P. & Ommer, B. High-resolution image synthesis with latent diffusion models. In Proc. IEEE/CVF Conference on Computer Vision and Pattern Recognition, 10684–10695 (IEEE, 2022).
- Dhariwal, P. & Nichol, A. Diffusion models beat GANs on image synthesis. Adv. Neural Inf. Process. Syst. 34, 8780–8794 (2021).
Google Scholar
- Ho, J. & Salimans, T. Classifier-free diffusion guidance. arXiv preprint https://doi.org/10.48550/arXiv.2207.12598 (2022).
- Chen, Z., Peng, B., Parthasarathy, S. & Ning, X. Shape-conditioned 3d molecule generation via equivariant diffusion models. arXiv preprint https://doi.org/10.48550/arXiv.2308.11890 (2023).
- Kwon, G. & Ye, J. C. Diffusion-based image translation using disentangled style and content representation. In International Conference on Learning Representations (ICLR, 2023).
- Preechakul, K., Chatthee, N., Wizadwongsa, S. & Suwajanakorn, S. Diffusion autoencoders: Toward a meaningful and decodable representation. In Proc. IEEE/CVF Conference on Computer Vision and Pattern Recognition, 10619–10629 (IEEE, 2022).
- Wang, Y. et al. Infodiffusion: representation learning using information maximizing diffusion models. In International Conference on Machine Learning, 36336–36354 (PMLR, 2023).
- Tolstikhin, I., Bousquet, O., Gelly, S. & Schoelkopf, B. Wasserstein auto-encoders. In International Conference on Learning Representations (ICLR, 2018).
- Axelrod, S. & Gomez-Bombarelli, R. Geom, energy-annotated molecular conformations for property prediction and molecular generation. Sci. Data 9, 185 (2022).
Google Scholar
- Landrum, G. et al. Rdkit: Open-source cheminformatics software (2016).
- Weininger, D. Smiles, a chemical language and information system. 1. introduction to methodology and encoding rules. J. Chem. Inf. Comput. Sci. 28, 31–36 (1988).
Google Scholar
- Lewis, P. et al. Retrieval-augmented generation for knowledge-intensive NLP tasks. Adv. Neural Inf. Process. Syst. 33, 9459–9474 (2020).
Google Scholar
- Wildman, S. A. & Crippen, G. M. Prediction of physicochemical parameters by atomic contributions. J. Chem. Inf. Comput. Sci. 39, 868–873 (1999).
Google Scholar
- Ertl, P. & Schuffenhauer, A. Estimation of synthetic accessibility score of drug-like molecules based on molecular complexity and fragment contributions. J. Cheminform. 1, 1–11 (2009).
Google Scholar
- Francoeur, P. G. et al. Three-dimensional convolutional neural networks and a cross-docked data set for structure-based drug design. J. Chem. Inf. Model. 60, 4200–4215 (2020).
Google Scholar
- Trott, O. & Olson, A. J. Autodock vina: improving the speed and accuracy of docking with a new scoring function, efficient optimization, and multithreading. J. Comput. Chem. 31, 455–461 (2010).
Google Scholar
- Alhossary, A., Handoko, S. D., Mu, Y. & Kwoh, C.-K. Fast, accurate, and reliable molecular docking with QuickVina 2. Bioinformatics 31, 2214–2216 (2015).
Google Scholar
- Huang, K. et al. Therapeutics data commons: machine learning datasets and tasks for drug discovery and development. In Proceedings of the Neural Information Processing Systems Track on Datasets and Benchmarks, Vol. 1 (2021).
- Guan, J. et al. 3D equivariant diffusion for target-aware molecule generation and affinity prediction. In International Conference on Learning Representations (ICLR, 2023).
- Luo, S., Guan, J., Ma, J. & Peng, J. A 3D generative model for structure-based drug design. Adv. Neural Inf. Process. Syst. 34, 6229–6239 (2021).
- Le, T., Cremer, J., Noé, F., Clevert, D.-A. & Schütt, K. Navigating the design space of equivariant diffusion-based generative models for de novo 3d molecule generation. In International Conference on Learning Representations (ICLR, 2024).
- Gretton, A., Borgwardt, K. M., Rasch, M. J., Schölkopf, B. & Smola, A. A kernel two-sample test. J. Mach. Learn. Res. 13, 723–773 (2012).
Google Scholar
- Gretton, A. et al. Optimal kernel choice for large-scale two-sample tests. Advances in Neural Information Processing Systems 25 (2012).
- Bowman, S. R. et al. Generating sentences from a continuous space. In Proceedings of the 20th SIGNLL Conference on Computational Natural Language Learning, 10–21 (Association for Computational Linguistics, 2016).
- Che, T., Li, Y., Jacob, A. P., Bengio, Y. & Li, W. Mode regularized generative adversarial networks. In International Conference on Learning Representations (ICLR, 2017).
- Satorras, V. G., Hoogeboom, E. & Welling, M. E (n) equivariant graph neural networks. In International Conference on Machine Learning, 9323–9332 (PMLR, 2021).
- Köhler, J., Klein, L. & Noé, F. Equivariant flows: exact likelihood generative learning for symmetric densities. In International Conference on Machine Learning, 5361–5370 (PMLR, 2020).
- Strang, G. Linear Algebra and Its Applications 2nd edn (Academic Press, Inc., 1980).
- Gaulton, A. et al. Chembl: a large-scale bioactivity database for drug discovery. Nucleic Acids Res. 40, D1100–D1107 (2012).
Google Scholar
- Scott, D. W. Multivariate Density Estimation: Theory, Practice, and Visualization (John Wiley & Sons, 2015).
- Li, T. Code release for “disentangled autoencoding equivariant diffusion model for controlled generation of 3D molecules” https://doi.org/10.5281/zenodo.18869528 (2026).
- Ochiai, T. et al. Variational autoencoder-based chemical latent space for large molecular structures with 3d complexity. Commun. Chem. 6, 249 (2023).
Google Scholar
- Lim, J., Ryu, S., Kim, J. W. & Kim, W. Y. Molecular generative model based on conditional variational autoencoder for de novo molecular design. J. Cheminform. 10, 31 (2018).
Google Scholar
- Romanelli, V. et al. Enhancing de novo drug design across multiple therapeutic targets with cvae generative models. ACS omega 9, 43963–43976 (2024).
Google Scholar
- Wei, L., Fu, N., Song, Y., Wang, Q. & Hu, J. Probabilistic generative transformer language models for generative design of molecules. J. Cheminform. 15, 88 (2023).
Google Scholar
- Sun, F., Zhan, Z., Guo, H., Zhang, M. & Tang, J. Graphvf: controllable protein-specific 3d molecule generation with variational flow. arXiv preprint https://doi.org/10.48550/arXiv.2304.12825 (2023).
Download references
Sign in to highlight and annotate this article

Conversation starters
Daily AI Digest
Get the top 5 AI stories delivered to your inbox every morning.
More about
model
AI Is Insatiable
While browsing our website a few weeks ago, I stumbled upon “ How and When the Memory Chip Shortage Will End ” by Senior Editor Samuel K. Moore. His analysis focuses on the current DRAM shortage caused by AI hyperscalers’ ravenous appetite for memory, a major constraint on the speed at which large language models run. Moore provides a clear explanation of the shortage, particularly for high bandwidth memory (HBM). As we and the rest of the tech media have documented, AI is a resource hog. AI electricity consumption could account for up to 12 percent of all U.S. power by 2028. Generative AI queries consumed 15 terawatt-hours in 2025 and are projected to consume 347 TWh by 2030. Water consumption for cooling AI data centers is predicted to double or even quadruple by 2028 compared to 2023. B

Anyone got Gemma 4 26B-A4B running on VLLM?
If yes, which quantized model are you using abe what’s your vllm serve command? I’ve been struggling getting that model up and running on my dgx spark gb10. I tried the intel int4 quant for the 31B and it seems to be working well but way too slow. Anyone have any luck with the 26B? submitted by /u/toughcentaur9018 [link] [comments]
Knowledge Map
Connected Articles — Knowledge Graph
This article is connected to other articles through shared AI topics and tags.
More in Models

Anyone got Gemma 4 26B-A4B running on VLLM?
If yes, which quantized model are you using abe what’s your vllm serve command? I’ve been struggling getting that model up and running on my dgx spark gb10. I tried the intel int4 quant for the 31B and it seems to be working well but way too slow. Anyone have any luck with the 26B? submitted by /u/toughcentaur9018 [link] [comments]

be careful on what could run on your gpus fellow cuda llmers
according to this report it seems that by "hammering" bits into dram chips through malicious cuda kernels, it could be possible to compromise systems equipped w/ several nvidia gpus up to excalating unsupervised privileged access to administrative role (root): https://arstechnica.com/security/2026/04/new-rowhammer-attacks-give-complete-control-of-machines-running-nvidia-gpus/ submitted by /u/DevelopmentBorn3978 [link] [comments]




Discussion
Sign in to join the discussion
No comments yet — be the first to share your thoughts!