CRISPR-GPT: AI gene-editing expert designed at Stanford

A long story short
CRISPR-GPT is an LLM-powered agent system developed by scientists from Stanford, Princeton and Google DeepMind, which automates gene-editing experiments, from CRISPR system selection to decision making when designing gene-editing RNA to data analysis.
The CRISPR-GPT project represents a unique collaboration between leading institutions in AI and biotechnology. Le Cong’s laboratory at Stanford University specializes in developing scalable genome editing and gene insertion technologies, integrating advances from metagenomics, computational biology and high-throughput engineering. The Princeton University team, led by Mengdi Wang and collaborator Denny Zhou at Google, brought expertise in machine learning and multi-agent systems. Together, they created a system that bridges the gap between cutting-edge AI and practical biological research, making gene editing accessible to scientists across disciplines.
Contents
Challenge: Making gene editing accessible to every scientist
Approach: Agentic AI development powered by Nebius infrastructure
Results: Democratizing gene editing through AI automation
Real-world validation with novice researchers
Superior performance over general AI
Takeaways: Specialized AI infrastructure drives scientific breakthroughs
Gene editing with CRISPR has already cured genetic diseases and ranks among the most-used laboratory techniques worldwide. However, designing effective gene-editing experiments requires deep expertise in molecular biology, specialized protocols and careful validation — knowledge that takes years to develop and limits who can harness this life-saving technology.
CRISPR-GPT changes everything. This groundbreaking collaboration between Le Cong’s research group at Stanford University, and collaborators from Princeton University and Google DeepMind, has created the first AI co-pilot that can guide any scientist through complete gene-editing workflows via natural conversation. The system handles everything from experimental planning and guide RNA design to protocol generation and data analysis, turning what once required months of expert training into accessible, automated processes.
Challenge: Making gene editing accessible to every scientist
CRISPR gene editing has transformed biological research and medicine, but its power comes with complexity that creates significant barriers to adoption. The technology requires expert-level decisions at every step — from selecting the right CRISPR system and designing guide RNAs, to choosing delivery methods and analyzing results.
Traditional gene editing demands deep understanding of molecular biology, specialized protocols and years of hands-on training. A typical PhD student needs weeks-to-months of intensive training just to become competent with basic techniques. This expertise bottleneck means that only specialized labs can fully leverage CRISPR’s potential, leaving countless researchers unable to apply gene editing to their important work.
Each gene editing experiment involves dozens of critical choices. Which Cas protein should be used? How should guide RNAs be designed for maximum efficiency and minimal off-target effects? What delivery method works best for the specific cell type? Making these decisions requires expertise across molecular biology, computational methods and experimental design.
Inconsistent results: Even experienced researchers struggle with reproducibility in gene editing experiments. Protocols that work in one lab may fail in another due to subtle differences. The high failure rate and time investment make gene editing risky for researchers working on tight timelines or with limited resources.
As CRISPR applications expand across drug discovery, cancer research, neurodegenerative disease studies, gene therapy, cell therapy and biological education, the shortage of trained experts becomes a critical bottleneck. Academic labs, biotech companies and pharmaceutical firms all compete for the same limited pool of gene editing specialists.
The challenge was clear: how to democratize access to this revolutionary technology while maintaining the precision and safety that gene editing requires.
Approach: Agentic AI development powered by Nebius infrastructure
To tackle these challenges, the research team developed CRISPR-GPT through a methodical, infrastructure-intensive approach that leveraged the advanced capabilities of Nebius AI Cloud.
Multi-agent architecture design
The team architected CRISPR-GPT as a multi-agent system with four core components: an LLM planner for task decomposition, task executors that implement complex workflows, user-proxy agents for natural conversation and tool providers that integrate external bioinformatics software. This modular design allows the system to handle the complex interdependencies inherent in gene editing workflows.
Accelerated development with Nebius AI Cloud
The team used Nebius AI Cloud’s GPU clusters to rapidly iterate on model architectures and fine-tuning approaches. Access to high-performance compute was crucial for training the specialized CRISPR-Llama3 model on 11 years of expert discussion data from scientific forums. The ability to quickly provision GPU resources enabled the team to test multiple model configurations in parallel, thus saving several months of development time.
Nebius’ flexible, scalable infrastructure allowed the researchers to transition seamlessly from small-scale prototyping to large-scale model training. The platform’s pre-configured AI libraries and drivers meant the team could immediately deploy their workloads, without spending time on infrastructure setup.
Comprehensive evaluation framework
The team developed Gene-editing bench, a comprehensive evaluation framework with 288 test cases covering experiment planning, guide RNA design, delivery method selection and Q&A scenarios. Human experts provided ground truth annotations, enabling rigorous benchmarking against existing tools and general-purpose LLMs.
Results: Democratizing gene editing through AI automation
The collaboration between the CRISPR-GPT research team and Nebius AI Cloud delivered breakthrough results that validate the potential for AI to democratize complex biotechnology.
Let us build pipelines of the same complexity for you
Our dedicated solution architects will examine all your specific requirements and build a solution tailored specifically for you.
Real-world validation with novice researchers
The most compelling proof of CRISPR-GPT’s effectiveness came from wet lab demonstrations with junior researchers who had no prior gene editing experience. In the first experiment, a PhD student used CRISPR-GPT to design and execute a multi-gene knockout targeting four genes (TGFβR1, SNAI1, BAX and BCL2L1) in human lung cancer cells. The AI guided every step from system selection to data analysis, achieving a consistent ~80% editing efficiency across all targets on the first attempt.
In a second experiment, an undergraduate student successfully performed epigenetic activation of two genes (NCR3LG1 and CEACAM1) in melanoma cells, achieving 56.5% and 90.2% efficiency, respectively. Both researchers succeeded without any prior training, demonstrating that CRISPR-GPT can indeed make expert-level gene editing accessible to any researcher.
Superior performance over general AI
Comprehensive benchmarking showed CRISPR-GPT consistently outperforming general-purpose LLMs like GPT-4o. In expert evaluations, CRISPR-GPT scored higher on accuracy, reasoning and action, completeness and conciseness across all major gene editing tasks. The system’s domain specialization eliminates common AI pitfalls like hallucination and irrelevant responses that plague general models in scientific contexts.
Takeaways: Specialized AI infrastructure drives scientific breakthroughs
Infrastructure as innovation catalyst
The CRISPR-GPT project demonstrates how access to high-performance AI infrastructure directly accelerates scientific breakthroughs. A single 8x NVIDIA H100 GPU node enabled rapid experimentation with models and training approaches that would have been impossible with limited compute resources. The ability to iterate quickly on complex multi-agent systems and fine-tune specialized models was crucial to the project’s success.
Domain specialization trumps general capability
While general-purpose LLMs have broad knowledge, specialized systems like CRISPR-GPT deliver dramatically superior performance in their target domains. The combination of domain-specific fine-tuning, expert-curated knowledge bases and purpose-built architectures creates AI systems that can truly compete with human experts. This suggests a future where specialized AI agents become standard tools across scientific disciplines.
Collaborative development unlocks potential
The success of CRISPR-GPT arose from deep collaboration between AI researchers, domain experts and infrastructure providers. Stanford’s biological expertise, Princeton’s AI capabilities, Google DeepMind’s technical resources and Nebius’ infrastructure created a synergy that no single organization could have achieved alone. This collaborative model may become essential for developing AI systems that can meaningfully impact complex scientific challenges.
Democratization through automation
By automating expert-level decision-making, CRISPR-GPT proves that AI can democratize access to powerful technologies, without compromising quality or safety. The system’s built-in safeguards, ethical guidelines and validated protocols ensure that novice researchers can achieve expert-level results while maintaining appropriate safety standards. This approach could transform how we think about training and expertise in rapidly advancing technical fields.
The CRISPR-GPT project represents more than just a successful AI application — it’s a blueprint for how specialized AI systems, powered by robust infrastructure and expert collaboration, can break down barriers and accelerate scientific progress. As gene editing applications expand across medicine, agriculture and biotechnology, tools like CRISPR-GPT will be essential for unleashing the full potential of these transformative technologies.
Acknowledgements
Yuanhao Qu (PhD Student, Stanford University, Le Cong Group) and Kaixuan Huang (PhD Student, Princeton University, Mengdi Wang Group) are co-first authors of this work. The Le Cong Group at Stanford University has been developing a series of AI agents for biomedical discovery in genomics. Beta testing for CRISPR-GPT is available by signing up at genomics.stanford.edu.
More exciting stories

vLLM
Using Nebius’ infrastructure, vLLM — a leading open-source LLM inference framework — is testing and optimizing their inference capabilities in different conditions, enabling high-performance, low-cost model serving in production environments.

SGLang
A pioneering LLM inference framework SGLang teamed up with Nebius AI Cloud to supercharge DeepSeek R1’s performance for real-world use. The SGLang team achieved a 2× boost in throughput and markedly lower latency on one node.

London Institute for Mathematical Sciences
How well can LLMs abstract problem-solving rules and how to test such ability? A research by LIMS, conducted using our compute, helps to understand the causes of LLM imperfections.