Published on

Can We Teach AI to Create Enzymes on Demand?

Scientists have developed EnzyControl, a system that «teaches» a computer to design enzymes for specific molecules – it’s like programming biological machines to perform desired tasks.

Biology & Neuroscience
DeepSeek-V3
Leonardo Phoenix 1.0
Author: Dr. Juan Mendoza Reading Time: 12 – 18 minutes

Accessible language

94%

Teaching talent

90%

Inspiring simplicity

93%
Original title: EnzyControl: Adding Functional and Substrate-Specific Control for Enzyme Backbone Generation
Publication date: Oct 29, 2025

Imagine you need to create the perfect key for a lock you've never seen. And this key must not just open the lock, but do so with incredible speed and precision, millions of times in a row, without wearing out. Sound like science fiction? This is precisely the challenge scientists face when designing new enzymes – the molecular machines that drive almost every chemical reaction in living organisms.

Why Create Enzymes in the First Place?

Enzymes are nature's catalysts, biological accelerators for chemical reactions. Without them, life as we know it would be impossible. Every second, billions of enzymes in your body are breaking down food, copying DNA, fighting toxins, and performing thousands of other tasks. They are so efficient they can speed up reactions by millions, or even billions, of times! 🚀

But what if we could learn to create our own enzymes for specific tasks? Imagine an enzyme that converts plastic waste into biofuel. Or a molecule capable of synthesizing a complex medicine in minutes instead of months of chemical synthesis. Or a biological «nanorobot» that destroys cancer cells while leaving healthy tissue untouched.

The problem is, creating such an enzyme is incredibly difficult. It’s like trying to write a program in a language with billions of letters, where the rules of grammar are defined by the laws of quantum physics and thermodynamics simultaneously.

Why Don't Conventional Methods Work?

Until recently, scientists tried to solve this problem in three ways. The first is a kind of «molecular selection»: take an existing enzyme and make random changes, hoping one of them improves its function. This method works, but it’s slow – like trying to improve a car by randomly swapping out bolts in the engine.

The second approach is computational modeling. We calculate how every atom in the enzyme interacts with the target molecule using the laws of physics. It's precise, but incredibly time-consuming. Simulating a single millisecond of a reaction can require weeks of calculations on a supercomputer.

The third path is creating enzymes «from scratch», or de novo design. We take a known active site – the part of the enzyme that directly performs the chemical reaction – and try to «pack» it into a new protein structure. This is like trying to fit a Ferrari engine into the body of a cargo van: theoretically possible, but in practice, you often end up with something non-functional.

The main problem with all these methods is that they miss the most important thing: the enzyme and its substrate (the molecule it works on) are a single, unified system. They must fit together like puzzle pieces. Moreover, it’s a 3D puzzle where it's not just the shape that matters, but also the distribution of electrical charges, the flexibility of the structure, and even how water molecules are arranged in the active site.

When Artificial Intelligence Learns Biology

In recent years, a revolution has taken place in the world of protein design. Neural networks have emerged that can predict the 3D structure of a protein from its amino acid sequence with an accuracy comparable to experimental methods. DeepMind's AlphaFold became for biology what the iPhone was for mobile communication – not the first, but the one that set the direction for decades to come.

Then came generative models – systems that don't just predict the structure of known proteins, but create entirely new ones. RFdiffusion, FrameFlow, Chroma – these systems can «dream up» proteins that have never existed in nature. Imagine a program that generates not images or text, but 3D molecular structures, potentially capable of life!

But when it comes to enzymes, even these advanced systems stumble. Why? Because they don't understand why the enzyme is being created. They can generate a beautiful, stable protein structure, but it will be useless if it can't efficiently bind to the right molecule and catalyze the necessary reaction.

It's like asking an architectural AI to design a building without explaining that it needs to be a concert hall with specific acoustics. You’ll get a beautiful building, but the music inside will sound terrible.

EnzyControl: When AI Starts to Understand Chemistry

A group of researchers decided to change the rules of the game. They created a system called EnzyControl, which works on a fundamentally different principle. Instead of generating an enzyme and then checking if it works with the desired substrate, the new system «knows» from the very beginning which molecule the enzyme is being created for.

How does it work? Let's use the architecture analogy again. Imagine you're designing a theater. The classic approach: first, draw a beautiful building, then try to cram a stage and auditorium inside. The EnzyControl approach: start with the requirements for acoustics, sightlines, and audience flow, and then build the entire structure of the building based on them.

At the heart of the system are three key components. The first is a base neural network that knows how to create protein structures around specified catalytic sites. These sites are like the building's foundation; they must stay in place because this is where the chemical magic happens.

The second component is a module called EnzyAdapter. This is a lightweight add-on to the main network that «weaves» information about the substrate into the generation process. Imagine a translator who doesn't just translate words but also explains the cultural context, making the communication truly effective. EnzyAdapter takes the chemical properties of the target molecule – its shape, charge distribution, flexibility – and translates this into a language the protein-generating system can understand.

The third element is a smart, two-stage training strategy. First, the system learns to recognize the connections between substrates and enzymes without changing the base model. This is like teaching an experienced chef to work with a new ingredient without re-teaching them the basic skills of cooking. Then, the entire system is fine-tuned, allowing it to create truly optimal structures.

The Database That Was Missing

But even the smartest neural network is useless without high-quality training data. It's like trying to teach someone to cook using only recipes from cooking shows, where half the ingredients are replaced with a «magic sauce» and the proportions are listed as «to taste» or «by eye».

The researchers created the EnzyBind database – a collection of over 11,000 «enzyme-substrate» pairs taken from real experimental structures. But this isn't just a set of molecular photographs. For each enzyme, they identified the conserved catalytic sites – the amino acid residues that evolution has preserved for millions of years because they are critical for function.

How was this done? The scientists used multiple sequence alignment – a technique that allows them to compare tens or hundreds of related enzymes and find which positions remain unchanged. It’s as if you collected hundreds of soup recipes from different grandmothers across the country and figured out which ingredients are present in absolutely all of them – those are what make up the essence of the dish.

This annotated data became the «textbook» for EnzyControl, allowing the system to learn not only how to create structures but also to understand which parts of those structures are functionally critical.

Results That Impress

Now for the most interesting part: does it work in practice? The researchers conducted a series of experiments, comparing EnzyControl with the best existing methods for enzyme design. The results were impressive.

The first metric was «designability». This is the proportion of created enzymes that can actually be synthesized and will fold into the correct structure. For EnzyControl, this figure was 71.6% – 13% higher than its closest competitor. For comparison: if before, six out of ten designed enzymes worked, now seven do. It may seem like a small improvement, but in the world of experimental biology, where synthesizing each candidate costs thousands of dollars and weeks of work, this is a huge breakthrough.

The second indicator was functional accuracy. Enzymes are classified by the EC (Enzyme Commission) nomenclature, which defines the specific chemical reaction an enzyme catalyzes. EnzyControl correctly «guessed» the enzyme class 50% of the time – 10% better than the alternatives. This means the system doesn't just create structurally correct proteins; it also understands their chemical function.

But most impressive of all were the catalytic efficiency metrics. The kcat constant, which describes how many substrate molecules an enzyme can process per second, was 15% higher for enzymes created by EnzyControl. The substrate binding energy improved by an average of 10-12%. These aren't just numbers – this is the difference between an enzyme that works in a lab and an enzyme that can become the basis for an industrial process.

When Less Means More

One of the unexpected results: the enzymes created by EnzyControl turned out to be, on average, 30% shorter than those from competing methods, while maintaining similar or even better catalytic activity. Why is this important?

In biotechnology, size matters. The shorter the protein, the easier and cheaper it is to synthesize. The bacteria or yeast used for production can produce more copies of a shorter protein. A more compact structure often means greater stability and resistance to denaturation.

This is reminiscent of the principle of elegant code in programming: the best code isn't the one that gets the job done by any means necessary, but the one that solves it with the minimum number of lines while remaining readable and efficient.

The Stress Test

But the real test of any machine learning system is its ability to work with data it has never seen during training. Can EnzyControl create an enzyme for a completely new substrate? Or for a class of reactions that wasn't in the training set?

The researchers conducted zero-shot tests – asking the system to create enzymes for molecules and reactions not found in the EnzyBind database. The results were encouraging: the average binding energy for these «unseen» cases was around $-7 \text{ kcal/mol}$, which is comparable to the training examples.

One case was particularly telling. For a known enzyme with the PDB identifier 2cv3, EnzyControl created an alternative version that bound to the substrate 51% more strongly and worked eight times more efficiently than the best result obtained by the RFDiffusion method. Eight times! Imagine if your car could suddenly drive eight times farther on the same amount of fuel.

What's Inside the Black Box?

One of the principles of good science is to understand why something works, not just that it works. The researchers conducted a series of tests, disabling various components of EnzyControl to understand which parts of the system were truly important.

When they removed the EnzyAdapter module – that «translator» between the substrate and the enzyme – the quality dropped sharply. Without it, the system reverted to creating structurally beautiful but functionally inefficient enzymes. This confirmed the key idea: the substrate must be part of the design process from the very beginning, not added as an afterthought.

When they turned off the catalytic site annotations derived from the multiple sequence alignments, the results were even worse. It turns out that information about evolutionarily conserved positions is critically important – it sets the functional «anchors» around which the rest of the structure is built.

These experiments showed that EnzyControl's success is not an accident or the result of simply increasing the model's size. It is a consequence of smart architecture, where every component plays a meaningful role.

Limitations and the Future

It would be dishonest not to mention the limitations. EnzyControl creates somewhat less diverse structures than some diffusion models. This is a trade-off: the system sacrifices creativity for functionality and reliability. Sometimes this is the right choice – it's better to have five working designs than twenty «creative» ones of which only one works.

Furthermore, the system currently only works with the types of chemical reactions and substrates that are well-represented in the database. This is a chicken-and-egg problem: to create enzymes for exotic reactions, you need examples of such enzymes, but if we had them, we might not need AI to design them.

Another limitation is computational. Although EnzyControl runs faster than full molecular modeling, creating and evaluating each enzyme still requires significant resources. Further optimizations are needed for widespread industrial application.

A Look to the Future

Nevertheless, EnzyControl represents an important step toward a future where we can design enzymes as routinely as we write software today. Imagine a world where:

Pharmaceutical companies create enzymes to synthesize complex drugs in days instead of years. A molecule that currently takes a decade and a billion dollars to develop could be synthesized biologically in months for a million dollars.

Biofuel plants use specialized enzymes to efficiently process agricultural waste into ethanol or diesel. Straw, wood chips, and even algae become raw materials for renewable energy.

Medicine gains personalized enzyme therapies. Your body can't break down a certain substance due to a genetic mutation? We'll create an enzyme for you that compensates for this defect, adapted to your unique biology.

The food industry uses enzymes to create products with enhanced flavors and nutritional properties without GMOs or chemical additives. Cheese without cows, meat without animals, all thanks to precision-designed biocatalysts.

Nature as a Hacker

Returning to my favorite idea: nature is the most brilliant hacker of all. Over three and a half billion years of evolution, it has sifted through an unimaginable number of variations, selecting and combining successful solutions. The enzymes working in your cells right now are the result of billions of iterations of nature's optimization algorithm.

But we have an advantage that evolution lacks: we can learn from its solutions, understand the principles underlying them, and apply them purposefully. We don't wait millions of years for random mutations – we use computational power, big data, and artificial intelligence to accelerate this process by trillions of times.

EnzyControl is not just another bioinformatics tool. It's a demonstration that we are beginning to speak nature's language fluently enough not only to read its code, but to write our own. We are learning to be co-authors of the great book of life, adding new chapters that nature itself might never have written.

And this is only the beginning of the journey. Every designed enzyme is a new page in this book, new proof that the line between the natural and the artificial is blurring. Perhaps in a few decades, our descendants will look at enzyme design as something commonplace, just as we view antibiotics or vaccines today. But for us, living in the era of these discoveries, every success is a small miracle, reminding us that science is not just formulas and experiments.

It is the ability to take the chaos of the molecular world and find order within it. It is the courage to assert that we can understand and improve even that which was created over billions of years. It is the infinite curiosity that drives us to ask not «why does this work»? but «how can we make it work better»?.

See you in the pages of the great book of life, which we are only just beginning to learn to read and write. 🧬

Original authors : Chao Song, Zhiyuan Liu, Han Huang, Liang Wang, Qiong Wang, Jianyu Shi, Hui Yu, Yihang Zhou, Yang Zhang
GPT-5
Claude Sonnet 4.5
Gemini 2.5 Pro
Previous Article How to Teach a Robot to Do Anything – Without a Single Lesson Next Article How to Teach an Algorithm Not to Panic: The Story of Prediction Intervals That Think Ahead

We believe in the power of human – AI dialogue

GetAtom was built so anyone can experience this collaboration first-hand: texts, images, and videos are just a click away.

Start today

+ get as a gift
100 atoms just for signing up

Lab

You might also like

Read more articles

When the Market Loses its Randomness: How Price Quirks Create Infinite Profit Opportunities

Research shows that in financial models with unusual price behavior – stops, reflections, asymmetry – strange arbitrage opportunities arise, resembling a «perpetual motion machine» of trading.

Finance & Economics

How Antennas Learned to Work Without Expensive Electronics: A Cylindrical Array for Future Networks

A new antenna architecture for 6G uses simple geometry instead of thousands of phase shifters – cutting costs by 15x while maintaining connection efficiency.

Electrical Engineering & System Sciences

When Geometry Sings: How Abstract Spaces Tell Stories Through Curves

Imagine spaces where shapes intertwine like musical notes, and counting them reveals invisible symmetries – this is the world of toric Calabi-Yau manifolds.

Mathematics & Statistics

Want to dive deeper into the
world of AI creations?

Be the first to discover new books, articles, and AI experiments on our Telegram channel!

Subscribe