As the capabilities of generative AI models have grown, you’ve probably seen how they can transform simple text prompts into hyperrealistic images and even extended video clips.
More recently, generative AI has shown potential in helping chemists and biologists explore static molecules, like proteins and DNA. Models like AlphaFold can predict molecular structures to accelerate drug discovery, and the MIT-assisted āRFdiffusion,ā for example, can help design new proteins. One challenge, though, is that molecules are constantly moving and jiggling, which is important to model when constructing new proteins and drugs. Simulating these motions on a computer using physics ā a technique known as molecular dynamics ā can be very expensive, requiring billions of time steps on supercomputers.
As a step toward simulating these behaviors more efficiently, MIT Computer Science and Artificial Intelligence Laboratory (CSAIL) and Department of Mathematics researchers have developed a generative model that learns from prior data. The teamās system, called MDGen, can take a frame of a 3D molecule and simulate what will happen next like a video, connect separate stills, and even fill in missing frames. By hitting the āplay buttonā on molecules, the tool could potentially help chemists design new molecules and closely study how well their drug prototypes for cancer and other diseases would interact with the molecular structure it intends to impact.
Co-lead author Bowen Jing SM ā22 says that MDGen is an early proof of concept, but it suggests the beginning of an exciting new research direction. āEarly on, generative AI models produced somewhat simple videos, like a person blinking or a dog wagging its tail,ā says Jing, a PhD student at CSAIL. āFast forward a few years, and now we have amazing models like Sora or Veo that can be useful in all sorts of interesting ways. We hope to instill a similar vision for the molecular world, where dynamics trajectories are the videos. For example, you can give the model the first and 10th frame, and itāll animate whatās in between, or it can remove noise from a molecular video and guess what was hidden.ā
The researchers say that MDGen represents a paradigm shift from previous comparable works with generative AI in a way that enables much broader use cases. Previous approaches were āautoregressive,ā meaning they relied on the previous still frame to build the next, starting from the very first frame to create a video sequence. In contrast, MDGen generates the frames in parallel with diffusion. This means MDGen can be used to, for example, connect frames at the endpoints, or āupsampleā a low frame-rate trajectory in addition to pressing play on the initial frame.
This work was presented in a paper shown at the Conference on Neural Information Processing Systems (NeurIPS) this past December. Last summer, it was awarded for its potential commercial impact at the International Conference on Machine Learningās ML4LMS Workshop.
Some small steps forward for molecular dynamics
In experiments, Jing and his colleagues found that MDGenās simulations were similar to running the physical simulations directly, while producing trajectories 10 to 100 times faster.
The team first tested their modelās ability to take in a 3D frame of a molecule and generate the next 100 nanoseconds. Their system pieced together successive 10-nanosecond blocks for these generations to reach that duration. The team found that MDGen was able to compete with the accuracy of a baseline model, while completing the video generation process in roughly a minute ā a mere fraction of the three hours that it took the baseline model to simulate the same dynamic.
When given the first and last frame of a one-nanosecond sequence, MDGen also modeled the steps in between. The researchersā system demonstrated a degree of realism in over 100,000 different predictions: It simulated more likely molecular trajectories than its baselines on clips shorter than 100 nanoseconds. In these tests, MDGen also indicated an ability to generalize on peptides it hadnāt seen before.
MDGenās capabilities also include simulating frames within frames, āupsamplingā the steps between each nanosecond to capture faster molecular phenomena more adequately. It can even āāāinpaintā structures of molecules, restoring information about them that was removed. These features could eventually be used by researchers to design proteins based on a specification of how different parts of the molecule should move.
Toying around with protein dynamics
Jing and co-lead author Hannes StƤrk say that MDGen is an early sign of progress toward generating molecular dynamics more efficiently. Still, they lack the data to make these models immediately impactful in designing drugs or molecules that induce the movements chemists will want to see in a target structure.
The researchers aim to scale MDGen from modeling molecules to predicting how proteins will change over time. āCurrently, weāre using toy systems,ā says StƤrk, also a PhD student at CSAIL. āTo enhance MDGenās predictive capabilities to model proteins, weāll need to build on the current architecture and data available. We donāt have a YouTube-scale repository for those types of simulations yet, so weāre hoping to develop a separate machine-learning method that can speed up the data collection process for our model.ā
For now, MDGen presents an encouraging path forward in modeling molecular changes invisible to the naked eye. Chemists could also use these simulations to delve deeper into the behavior of medicine prototypes for diseases like cancer or tuberculosis.
āMachine learning methods that learn from physical simulation represent a burgeoning new frontier in AI for science,ā says Bonnie Berger, MIT Simons Professor of Mathematics, CSAIL principal investigator, and senior author on the paper. āMDGen is a versatile, multipurpose modeling framework that connects these two domains, and weāre very excited to share our early models in this direction.ā
āSampling realistic transition paths between molecular states is a major challenge,ā says fellow senior author Tommi Jaakkola, who is the MIT Thomas Siebel Professor of electrical engineering and computer science and the Institute for Data, Systems, and Society, and a CSAIL principal investigator. āThis early work shows how we might begin to address such challenges by shifting generative modeling to full simulation runs.ā
Researchers across the field of bioinformatics have heralded this system for its ability to simulate molecular transformations. āMDGen models molecular dynamics simulations as a joint distribution of structural embeddings, capturing molecular movements between discrete time steps,ā says Chalmers University of Technology associate professor Simon Olsson, who wasnāt involved in the research. āLeveraging a masked learning objective, MDGen enables innovative use cases such as transition path sampling, drawing analogies to inpainting trajectories connecting metastable phases.ā
The researchersā work on MDGen was supported, in part, by the National Institute of General Medical Sciences, the U.S. Department of Energy, the National Science Foundation, the Machine Learning for Pharmaceutical Discovery and Synthesis Consortium,Ā the Abdul Latif Jameel Clinic for Machine Learning in Health, the Defense Threat Reduction Agency, and the Defense Advanced Research Projects Agency.