MIT scientists debut a generative AI model that could create molecules addressing hard-to-treat diseases
BoltzGen generates protein binders for any biological target from scratch, expanding AI’s reach from understanding biology toward engineering it.
More than 300 people across academia and industry spilled into an auditorium to attend a BoltzGen seminar on Thursday, Oct. 30, hosted by the Abdul Latif Jameel Clinic for Machine Learning in Health (MIT Jameel Clinic). Headlining the event was MIT PhD student and BoltzGen’s first author Hannes Stärk, who had announced BoltzGen just a few days prior.
Building upon Boltz-2, an open-source biomolecular structure prediction model predicting protein binding affinity that made waves over the summer, BoltzGen (officially released on Sunday, Oct. 26.) is the first model of its kind to go a step further by generating novel protein binders that are ready to enter the drug discovery pipeline.
Three key innovations make this possible: first, BoltzGen’s ability to carry out a variety of tasks, unifying protein design and structure prediction while maintaining state-of-the-art performance. Next, BoltzGen’s built-in constraints are designed with feedback from wetlab collaborators to ensure the model creates functional proteins that don’t defy the laws of physics or chemistry. Lastly, a rigorous evaluation process tests the model on “undruggable” disease targets, pushing the limits of BoltzGen’s binder generation capabilities.
Most models used in industry or academia are capable of either structure prediction or protein design. Moreover, they’re limited to generating certain types of proteins that bind successfully to easy “targets.” Much like students responding to a test question that looks like their homework, as long as the training data looks similar to the target during binder design, the models often work. But existing methods are nearly always evaluated on targets for which structures with binders already exist, and end up faltering in performance when used on more challenging targets.
“There have been models trying to tackle binder design, but the problem is that these models are modality-specific,” Stärk points out. “A general model does not only mean that we can address more tasks. Additionally, we obtain a better model for the individual task since emulating physics is learned by example, and with a more general training scheme, we provide more such examples containing generalizable physical patterns.”
The BoltzGen researchers went out of their way to test BoltzGen on 26 targets, ranging from therapeutically relevant cases to ones explicitly chosen for their dissimilarity to the training data.
This comprehensive validation process, which took place in eight wetlabs across academia and industry, demonstrates the model’s breadth and potential for breakthrough drug development.
Parabilis Medicines, one of the industry collaborators that tested BoltzGen in a wetlab setting, praised BoltzGen’s potential: “we feel that adopting BoltzGen into our existing Helicon peptide computational platform capabilities promises to accelerate our progress to deliver transformational drugs against major human diseases.”
While the open-source releases of Boltz-1, Boltz-2, and now BoltzGen (which was previewed at the 7th Molecular Machine Learning Conference on Oct. 22) bring new opportunities and transparency in drug development, they also signal that biotech and pharmaceutical industries may need to reevaluate their offerings.
Amid the buzz for BoltzGen on the social media platform X, Justin Grace, a principal machine learning scientist at LabGenius, raised a question. “The private-to-open performance time lag for chat AI systems is [seven] months and falling,” Grace wrote in a post. “It looks to be even shorter in the protein space. How will binder-as-a-service co’s be able to [recoup] investment when we can just wait a few months for the free version?”
For those in academia, BoltzGen represents an expansion and acceleration of scientific possibility. “A question that my students often ask me is, ‘where can AI change the therapeutics game?’” says senior co-author and MIT Professor Regina Barzilay, AI faculty lead for the Jameel Clinic and an affiliate of the Computer Science and Artificial Intelligence Laboratory (CSAIL). “Unless we identify undruggable targets and propose a solution, we won’t be changing the game,” she adds. “The emphasis here is on unsolved problems, which distinguishes Hannes’ work from others in the field.”
Senior co-author Tommi Jaakkola, the Thomas Siebel Professor of Electrical Engineering and Computer Science who is affiliated with the Jameel Clinic and CSAIL, notes that "models such as BoltzGen that are released fully open-source enable broader community-wide efforts to accelerate drug design capabilities.”
Looking ahead, Stärk believes that the future of biomolecular design will be upended by AI models. “I want to build tools that help us manipulate biology to solve disease, or perform tasks with molecular machines that we have not even imagined yet,” he says. “I want to provide these tools and enable biologists to imagine things that they have not even thought of before.”
Sign in to highlight and annotate this article

Conversation starters
Daily AI Digest
Get the top 5 AI stories delivered to your inbox every morning.
More about
modelTraining mRNA Language Models Across 25 Species for $165
<table> <tr><td> <a href="https://www.reddit.com/r/LocalLLaMA/comments/1s912ur/training_mrna_language_models_across_25_species/"> <img src="https://external-preview.redd.it/nVu8HcAwsZ22nurNc_vAW-R_2drotHTdQavTOwGgmi4.png?width=640&crop=smart&auto=webp&s=00c766290f7df67dcd9c20ff789b8f25a14b886b" alt="Training mRNA Language Models Across 25 Species for $165" title="Training mRNA Language Models Across 25 Species for $165" /> </a> </td><td> <!-- SC_OFF --><div class="md"><p>We built an end-to-end protein AI pipeline covering structure prediction, sequence design, and codon optimization. After comparing multiple transformer architectures for codon-level language modeling, CodonRoBERTa-large-v2 emerged as the clear winner with a perplexity of 4.10 and a Spearman CAI correlation of 0
[llama.cpp] New TurboQuant 3-bit KV Cache is insane! 17 t/s on Nemotron 30B using only 8GB VRAM (Full Windows/MSVC Build Guide + Auto-Script)
<table> <tr><td> <a href="https://www.reddit.com/r/LocalLLaMA/comments/1s931oz/llamacpp_new_turboquant_3bit_kv_cache_is_insane/"> <img src="https://preview.redd.it/ul2pi1esogsg1.jpg?width=140&height=74&auto=webp&s=37edd20c35b0338dbf6b160a01342c2f88b2eb90" alt="[llama.cpp] New TurboQuant 3-bit KV Cache is insane! 17 t/s on Nemotron 30B using only 8GB VRAM (Full Windows/MSVC Build Guide + Auto-Script)" title="[llama.cpp] New TurboQuant 3-bit KV Cache is insane! 17 t/s on Nemotron 30B using only 8GB VRAM (Full Windows/MSVC Build Guide + Auto-Script)" /> </a> </td><td> <!-- SC_OFF --><div class="md"><p><a href="https://preview.redd.it/ul2pi1esogsg1.jpg?width=707&format=pjpg&auto=webp&s=20d2b478c1700f74ce0933f9202177f07c19239e"> </a></p> <h1>Hi everyone.</h1> <p>If you a
You guys seen this? 1-bit model with an MMLU-R of 65.7, 8B params
<!-- SC_OFF --><div class="md"><p>This is nuts.</p> <p><a href="https://huggingface.co/prism-ml/Bonsai-8B-gguf">prism-ml/Bonsai-8B-gguf · Hugging Face</a></p> <p>has anyone tested this thing?</p> </div><!-- SC_ON -->   submitted by   <a href="https://www.reddit.com/user/OmarBessa"> /u/OmarBessa </a> <br/> <span><a href="https://www.reddit.com/r/LocalLLaMA/comments/1s91jxl/you_guys_seen_this_1bit_model_with_an_mmlur_of/">[link]</a></span>   <span><a href="https://www.reddit.com/r/LocalLLaMA/comments/1s91jxl/you_guys_seen_this_1bit_model_with_an_mmlur_of/">[comments]</a></span>
Knowledge Map
Connected Articles — Knowledge Graph
This article is connected to other articles through shared AI topics and tags.
More in Models
Training mRNA Language Models Across 25 Species for $165
<table> <tr><td> <a href="https://www.reddit.com/r/LocalLLaMA/comments/1s912ur/training_mrna_language_models_across_25_species/"> <img src="https://external-preview.redd.it/nVu8HcAwsZ22nurNc_vAW-R_2drotHTdQavTOwGgmi4.png?width=640&crop=smart&auto=webp&s=00c766290f7df67dcd9c20ff789b8f25a14b886b" alt="Training mRNA Language Models Across 25 Species for $165" title="Training mRNA Language Models Across 25 Species for $165" /> </a> </td><td> <!-- SC_OFF --><div class="md"><p>We built an end-to-end protein AI pipeline covering structure prediction, sequence design, and codon optimization. After comparing multiple transformer architectures for codon-level language modeling, CodonRoBERTa-large-v2 emerged as the clear winner with a perplexity of 4.10 and a Spearman CAI correlation of 0
You guys seen this? 1-bit model with an MMLU-R of 65.7, 8B params
<!-- SC_OFF --><div class="md"><p>This is nuts.</p> <p><a href="https://huggingface.co/prism-ml/Bonsai-8B-gguf">prism-ml/Bonsai-8B-gguf · Hugging Face</a></p> <p>has anyone tested this thing?</p> </div><!-- SC_ON -->   submitted by   <a href="https://www.reddit.com/user/OmarBessa"> /u/OmarBessa </a> <br/> <span><a href="https://www.reddit.com/r/LocalLLaMA/comments/1s91jxl/you_guys_seen_this_1bit_model_with_an_mmlur_of/">[link]</a></span>   <span><a href="https://www.reddit.com/r/LocalLLaMA/comments/1s91jxl/you_guys_seen_this_1bit_model_with_an_mmlur_of/">[comments]</a></span>
[P] Looking for people who have had training runs fail unexpectedly to beta test a stability monitor. Free, takes 5 minutes to add to your existing loop. DM me.
<!-- SC_OFF --><div class="md"><p>Anyone actively training models want to try a stability monitor on a real run? Trying to get real world validation outside my own benchmarks.</p> </div><!-- SC_ON -->   submitted by   <a href="https://www.reddit.com/user/Turbulent-Tap6723"> /u/Turbulent-Tap6723 </a> <br/> <span><a href="https://www.reddit.com/r/MachineLearning/comments/1s93kzm/p_looking_for_people_who_have_had_training_runs/">[link]</a></span>   <span><a href="https://www.reddit.com/r/MachineLearning/comments/1s93kzm/p_looking_for_people_who_have_had_training_runs/">[comments]</a></span>
FOR ME, Qwen3.5-27B is better than Gemini 3.1 Pro and GPT-5.3 Codex
<!-- SC_OFF --><div class="md"><p>There's something I hate about the big SOTA proprietary models. In order to make them better for people who don't know how to program, they're optimized to solve problems entirely autonomously. Yeah, this makes people over on <a href="/r/ChatGPT">/r/ChatGPT</a> soypog when it writes a 7z parser in Python because the binary is missing, however, for me, this makes them suck. If something isn't matching up, Qwen3.5-27B will just give up. If you're trying to vibecode some slop this is annoying, but for me this is much, much better. I'm forced to use GitHub Copilot in university, and whenever there's a problem, it goes completely off the rails and does some absolute hogwash. Like, for example, it was struggling to write to a file that had some broken permission

Discussion
Sign in to join the discussion
No comments yet — be the first to share your thoughts!