Bangkok Post

CREATING PROTEINS WITH AI

Using digital art generators, scientists are testing new structures to fight cancer, flu and Covid

- CADE METZ © 2023 THE NEW YORK TIMES COMPANY

Last year, an artificial intelligen­ce lab called OpenAI unveiled technology that lets you create digital images simply by describing what you want to see. Called DALL-E, it sparked a wave of similar tools with names like Midjourney and Stable Diffusion. Promising to speed the work of digital artists, this new breed of AI captured the imaginatio­n of both the public and the pundits — and threatened to generate new levels of online disinforma­tion.

Social media is now teeming with the surprising­ly conceptual, in which shockingly detailed, often photoreali­stic images are generated by DALL-E and other tools. “Photo of a teddy bear riding a skateboard in Times Square”, “Cute corgi in a house made out of sushi”, “Jeflon Zuckergate­s”.

But when some scientists consider this technology, they see more than just a way of creating fake photos. They see a path to a new cancer treatment or a new flu vaccine or a new pill that helps you digest gluten.

Using many of the same techniques that underpin DALL-E and other art generators, these scientists are generating blueprints for new proteins — tiny biological mechanisms that can change the way our bodies behave.

Our bodies naturally produce about 20,000 proteins, which handle everything from digesting food to moving oxygen through the bloodstrea­m. Now, researcher­s are working to create proteins that are not found in nature, hoping to improve our ability to fight disease and do things that our bodies cannot do on their own.

David Baker, the director of the Institute for Protein Design at the University of Washington, has been working to build artisanal proteins for more than 30 years. By 2017, he and his team had shown this was possible. But they did not anticipate how the rise of new AI technologi­es would suddenly accelerate this work, shrinking the time needed to generate new blueprints from years down to weeks.

“What we need are new proteins that can solve modern-day problems, like cancer and viral pandemics,” Baker said. “We can’t wait for evolution.” He added, “Now, we can design these proteins much faster, and with much higher success rates, and create much more sophistica­ted molecules that can help solve these problems.”

Last year, Baker and his fellow researcher­s published a pair of papers in the journal Science describing how various AI techniques could accelerate protein design. But these papers have already been eclipsed by a newer one that draws on the techniques that drive tools like DALL-E, showing how new proteins can be generated from scratch much like digital photos.

“One of the most powerful things about this technology is that, like DALLE, it does what you tell it to do,” said Nate Bennett, one of the researcher­s working in the University of Washington lab. “From a single prompt, it can generate an endless number of designs.”

To generate images, DALL-E relies on what AI researcher­s call a neural network, a mathematic­al system loosely modelled on the network of neurons in the brain. This is the same technology that recognises the commands you bark into your smartphone, enables self-driving cars to identify (and avoid) pedestrian­s and translates languages on services like Skype.

A neural network learns skills by analysing vast amounts of digital data. By pinpointin­g patterns in thousands of corgi photos, for instance, it can learn to recognise a corgi. With DALLE, researcher­s built a neural network that looked for patterns as it analysed millions of digital images and the text captions that described what each of these images depicted. In this way, it learned to recognise the links between the images and the words.

When you describe an image for DALL-E, a neural network generates a set of key features that this image may include. One feature might be the curve of a teddy bear’s ear. Another might be the line at the edge of a skateboard. Then, a second neural network — called a diffusion model — generates the pixels needed to realise these features.

The diffusion model is trained on a series of images in which noise — imperfecti­on — is gradually added to a photograph until it becomes a sea of random pixels. As it analyses these images, the model learns to run this process in reverse. When you feed it random pixels, it removes the noise, transformi­ng these pixels into a coherent image.

At the University of Washington, other academic labs and new startups, researcher­s are using similar techniques in their effort to create new proteins.

Proteins begin as strings of chemical compounds, which then twist and fold into three-dimensiona­l shapes that define how they behave. In recent years, AI labs like DeepMind, owned by Alphabet, the same parent company as Google, have shown that neural networks can accurately guess the three-dimensiona­l shape of any protein in the body based just on the smaller compounds it contains — an enormous scientific advance.

Now, researcher­s like Baker are taking another step, using these systems to generate blueprints for entirely new proteins that do not exist in nature. The goal is to create proteins that take on very specific shapes; a particular shape can serve a particular task, such as fighting the virus that causes Covid-19.

Much as DALL-E leverages the relationsh­ip between captions and photograph­s, similar systems can leverage the relationsh­ip between a descriptio­n of what the protein can do and the shape it adopts. Researcher­s can provide a rough outline for the protein they want, then a diffusion model can generate its three-dimensiona­l shape.

The difference is that the human eye can instantly judge the fidelity of a DALL-E image. It cannot do the same with a protein structure. After AI technologi­es produce these protein blueprints, scientists must still take them into a wet lab — where experiment­s can be done with real chemical compounds — and make sure they do what they are supposed to do.

For this reason, some experts say that the latest AI technologi­es should be taken with a grain of salt. “Making a new structure is just a game,” said Frances Arnold, a Nobel Laureate who is a professor specialisi­ng in protein engineerin­g at the California Institute of Technology. “What really matters is: What can that structure actually do?”

From a single prompt, it can generate an endless number of designs

But for many researcher­s, these new techniques are not just accelerati­ng the creation of new protein candidates for the wet lab. They provide a way of exploring new innovation­s that researcher­s could not previously explore on their own.

“What’s exciting isn’t just that they are creative and explore unexpected possibilit­ies, but that they are creative while satisfying certain design objectives or constraint­s,” said Jue Wang, a researcher at the University of Washington. “This saves you from needing to check every possible protein in the universe.”

Often, artificial­ly intelligen­t machines are developed to perform skills that come naturally to humans, like piecing together images, writing text or playing board games. Protein-designing bots pose a more profound question, Wang said: “What can machines do that humans can’t do at all?”

 ?? ?? ABOVE
Namrata Anand, a former Stanford University researcher who is building a company in generative artificial intelligen­ce protein design.
ABOVE Namrata Anand, a former Stanford University researcher who is building a company in generative artificial intelligen­ce protein design.
 ?? ?? David Baker, the director of the Institute for Protein Design at the University of Washington.
David Baker, the director of the Institute for Protein Design at the University of Washington.
 ?? ?? A model of an artificial intelligen­ce-generated protein.
A model of an artificial intelligen­ce-generated protein.

Newspapers in English

Newspapers from Thailand