November Ultra-fast protein structure prediction to capture effects of sequence variation in mutation movies Konstantin Weissenow k.weissenow@tum.de 2 3 Michael Heinzinger 2 3 Martin Steinegger 1 Burkhard Rost 0 2 Institute for Advanced Study (TUM-IAS), Lichtenbergstr. 2a, 85748 Garching/Munich, Germany & TUM School of Life Sciences Weihenstephan (WZW) , Alte Akademie 8, Freising , Germany & Institute for Advanced Study (TUM- IAS) , Lichtenbergstr. 2a, 85748 Garching/Munich , Germany School of Biological Sciences, Seoul National University , Seoul , South Korea & Campus-Institute Data Science (CIDAS) , Goldschmidtstrasse 1, 37077 Göttingen , Germany & Artificial Intelligence Institute, Seoul National University , Seoul , South Korea & Institute of Molecular Biology and Genetics, Seoul National University , Seoul , South Korea TUM (Technical University of Munich) Department of Informatics, Bioinformatics & Computational Biology - i12 , Boltzmannstr. 3, 85748 Garching/Munich , Germany TUM Graduate School, Center of Doctoral Studies in Informatics and its Applications (CeDoSIA) , Boltzmannstr. 11, 85748 Garching , Germany 2022 18 2022

Top protein three-dimensional (3D) structure predictions require evolutionary information from multiple-sequence alignments (MSAs) and deep, convolutional neural networks and appear insensitive to small sequence changes. Here, we describe EMBER3D using embeddings from the pre-trained protein language model (pLM) ProtT5 to predict 3D structure directly from single sequences. Orders of magnitude faster than others, EMBER3D predicts average-length structures in milliseconds on consumer-grade machines. Although not nearly as accurate as AlphaFold2 , the speed of EMBER3D allows a glimpse at future applications such as the almost real-time rendering of deep mutational scanning (DMS) movies that visualize the effect of all point mutants on predicted structures. This also enables live-editing of sequence/structure pairs. EMBER3D is accurate enough for highly sensitive rapid remote homology detection by Foldseek identifying structural similarities. Overall, our use cases suggest that speed can complement accuracy, in particular when accessible through consumer-grade machines. EMBER3D is free and publicly available: https://github.com/kWeissenow/EMBER3D .

protein structure prediction deep learning machine learning protein language model transfer learning
-

Abbreviations used: 2D, two-dimensional; 2D structure: inter-residue distances/contacts; 3D, threedimensional; 3D structure: coordinates of atoms in a protein structure; AH, attention heads; BFD: big fantastic database; CASP, Critical Assessment of protein Structure Prediction; CNN, convolutional neural network; DCA, direct coupling analysis; DL, Deep Learning; DMS: deep mutational scanning; ML, machine learning; MSA, multiple sequence alignment; pLM, protein Language Model; PMM, protein mutation movie; SAV: single amino-acid variant; SCOP: Structural Classification of Proteins; SOTA, state-of-the-art.

Introduction

AlphaFold2 advances protein structure prediction. The Critical Assessment of protein Structure Prediction (CASP) has provided the gold-standard to evaluate protein structure prediction for almost three decades 1. At its first meeting (CASP1 Dec. 1994), the combination of machine learning (ML) and evolutionary information derived from multiple sequence alignments (MSAs) reported a major breakthrough in secondary structure prediction 2. 2021’s method of the year 3, AlphaFold2 4 has combined more advanced machine learning (ML) with larger MSAs and more potent hardware to substantially advance protein structure prediction. But even this pinnacle of 50 years of research has a shortcoming: predictions are mostly insensitive to small variations in the input 5. Protein language models (pLMs) substitute evolutionary information. Advances in natural language processing (NLP) spawned protein language models (pLMs) 6-11 that leverage the wealth of information in exponentially growing but unlabelled protein sequence databases by solely relying on sequential patterns found in the input. Processing the information learned by such pLMs, e.g., by inputting a protein sequence into the network and constructing vectors from the activation in the network’s last layers, yields a representation of protein sequences referred to as embeddings (Fig. S1 11). This allows to transfer features learned by the pLM to any downstream (prediction) task requiring numerical protein representations (transfer learning) which has already been showcased for various aspects of protein prediction ranging from structure 5,12 and disorder 13 to function 14. Distance in embedding space correlates more with protein function than with sequence similarity 15 and can help clustering proteins into families 16-18. Recently, pLMs have been used to directly predict protein 3D structure 5,19-22. Using embeddings from pLMs instead of evolutionary information from MSAs simplifies and speeds up structure prediction. Speed is gained at the price of accuracy while precomputed AlphaFold2 predictions are available for over 200 million proteins 23. Is there a value in gaining speed even when losing accuracy? Here, we introduce a novel solution using sequence embeddings and attention heads (AHs) from ProtT5 11 to predict inter-residue distances (2D structure) and backbone atom coordinates (3D structure). Without any MSA, our proposed solution reaches competitive performance at unprecedented speed. Paired with its sensitivity to small changes in the input, we showed that the proposed solution opens the door for novel applications such as the generation of <protein mutation movies = (PMMs). Each frame in these movies shows the predicted structure of a mutant. Connecting these frames, the rendered movie visualizes the mutational landscape of a protein. We also showed that our prediction quality sufficed for fast structure alignment, albeit not reaching the performance of AlphaFold2-like solutions.

Ultra-fast: protein structure predicted in sub-seconds . We benchmarked the prediction speed of EMBER3D through poly-alanine sequences with 20-1500 residues (i.e., alanines). EMBER3D predicted backbone atom coordinates (3D structure) and inter-residue distances (2D structure) in sub-seconds for average length sequences (Fig. 1B, e.g., 0.3s for 384 residues). In contrast, ESMFold 21 needs 14.2s (factor 47) for the same sequence, i.e. EMBER3D was as much faster than ESMFold as Results that sped up over AlphaFold2. Additionally, our model is small enough to fit onto consumer-grade GPU hardware (e.g., NVIDIA Geforce 1080ti with 8 GB of video memory) which can predict for most proteins (<430 residues). On server-grade hardware (e.g., NVIDIA Quadro RTX 8000 with 48 GB), proteins of up to ~700 residues are predicted in less than a second while the maximal length rises to 1420 residues (Fig. 1B). To put this into perspective, we applied the publicly available implementation ColabFold 24 speeding up AlphaFold2 4 and two newer 3D prediction methods using pLMs, namely OmegaFold 20 and HelixFold 19 on our test set (CASP14 domains 25) using the same hardware (NVIDIA Quadro RTX 8000). EMBER3D significantly outpaced the others (Fig. 1A). Its speed allowed predicting structures for 538,488 proteins from the entire Swiss-Prot database 26 in just 60.5 hours on a single server-grade GPU. Since structural prediction is obtained nearinstantaneously our solution allows live-editing of sequences.

Single-sequence predictions better than for AlphaFold2 . In order to assess the quality of EMBER3D predictions, we compared TM-score and lDDT27 on our test set (CASP14 domains) with results from ESMFold, OmegaFold, HelixFold, our previous method, EMBER2, and AlphaFold2 with MSAs (standard) and without using MSAs as input (Fig. 1CD). Although AlphaFold2/ColabFold significantly outperformed the EMBER family, performance dropped substantially without MSA input (Fig. 1CD: <AlphaFold2 single=). While HelixFold performed similar to the EMBER family, OmegaFold outperformed all sequence-based methods, including ESMFold, albeit at significantly higher runtime (Fig. 1A).

EMBER family more sensitive to mutants (SAVs) than AlphaFold2 . In contrast to methods relying on MSAs, we hypothesized our single sequence-based method to be more sensitive to small sequence variations, e.g., to single amino acid variants (SAVs ). We measured the structural effect of SAVs by computing the changes in structures predicted between wild type and mutated sequence measured in lDDT. Since EMBER3D outputs 2D and 3D structure, we considered both. We correlated predicted structural changes with deep mutational scanning (DMS) data for nine proteins 28-34 shorter than 250 residues. For AlphaFold2, our restricted resources forced to consider only the shortest five (loading our hardware for over a week). The SAV analysis required predicting 3D structures for all possible point mutants in a protein (e.g. 6,650 predictions for a protein of 350 residues). EMBER3D, completed this job within minutes (almost 10,000 times faster than ColabFold which is faster than AlphaFold2).

The differences between native and mutant structure predicted by AlphaFold2 correlated only weakly with DMS data (correlation over 5 proteins: 0.15, Fig. 2), significantly less than differences predicted by EMBER3D (Fig. 2). Amongst the EMBER3D methods, differences in predicted distance maps correlated more (0.36) than the 3D output (0.24, Fig. 2). While pLM-based ESMFold correlated better with DMS than AlphaFold2 (0.18), it did not reach EMBER3D on any sample, and OmegaFold remained even below (correlation over 5: 0.05). Only HelixFold outperformed the EMBER-family for one sample (Ubiquitin), despite a low overall correlation (0.18). Family-averaged AlphaFold2 and the protein-specific EMBER -family differed on average and in detail, e.g., the DMS set predicted by far best by AlphaFold2 (SUMO-conjugating enzyme UBC9 ) was below average for EMBER3D. Protein mutation movies capture differences in predicted structures. To visualize the mutational landscape of a protein, we developed a tool which renders the differences between structures predicted for wild type and all possible point mutants (SAVs: length*19) into a protein mutation movie (PMM). This tool first runs EMBER3D to obtain both 2D and 3D structure predictions for all possible SAVs in a protein sequence. Next, the tool generates three images for each mutant (SAV): ( 1 ) distance map for SAV (using matplotlib 35; Fig. 3C), ( 2 ) sketch of 3D backbone coordinates for SAV from a static viewpoint (using PyMol 36; Fig. 3B), and ( 3 ) a mutation profile (Fig. 3D). The tool then composes these three into one image (for each mutant), and renders the individual images as an animated movie clip using ffmpeg 37 with 19 frames per second, thereby showing all 19 possible amino-acid substitutions for one residue position in each second of playback time (Fig. 3). Predicted confidence correlates between EMBER3D and AlphaFold2 . Copying ideas introduced by AlphaFold2 4, EMBER3D also predicted its own reliability. Despite reaching lower overall performance than AlphaFold2, EMBER3D succeeded in assessing its own success (Fig. S2A). Although different in methodology and focus, we expected AlphaFold2 and EMBER3D to share the same trend in their ability to accurately predict structures for difficult proteins. Indeed, the predicted lDDT (pLDDT) confidence scores for 538,488 proteins from Swiss-Prot Spearman rank correlated to 0.53 (p-value ~0.0).

EMBER3D, optimized for speed, was overall less confident than AlphaFold2 in its own predictions. However, most proteins for which AlphaFold2 predicted a high-quality structure were also predicted with higher-than-average confidence by EMBER3D (Fig. S2B). This suggested using EMBER3D as a rough, but very fast pre-for AlphaFold2.

Foldseek paired with EMBER3D beats MMSeqs2 . Protein structure predictions can unravel relations between proteins with highly diverged sequences38. We benchmarked the detection of SCOP folds from EMBER3D predictions by predicting structures for 11,211 domains from SCOP-40 (v2.01) 39 and aligning these all against all with Foldseek 40. Predicted structures constituted the queries, experimental structures the inference lookup set 40,41. Replicating the Foldseek benchmark 40, we evaluated performance by measuring the sensitivity up to the 5th false positive per query for the three different levels of family , superfamily , and fold . Protein pairs with experimental structures were detected best by Foldseek (Fig. 4: sensitivity: 0.901, 0.578, 0.154 for family, superfamily and fold, respectively) followed by comparing EMBER3D-predicted pairs of structures (Fig. 4: 0.865, 0.499, 0.100), while sequence-sequence comparisons with MMseqs2 were substantially less successful (Fig. 4: 0.543, 0.082, 0.002).

Discussion

Speed enables mutation movies for higher order mutants. AlphaFold2 4 and RoseTTAFold 42 continue to mark the state-of-the-art in protein structure prediction by using evolutionary information derived from multiple sequence alignments (MSAs). ColabFold 24 speeds up the sequence search without losing performance. HelixFold 19, OmegaFold 20, ESMFold 21, and RGN2 22 speed up by replacing evolutionary information with embeddings from protein Language Models (pLMs). EMBER3D pushes much further: where ESMFold speeds up 10-60 times, EMBER3D adds more than another order of magnitude, even on less advanced hardware. Still, through its speed-up, ESMFold tripled the number of predictions to cover 600 million proteins 23,43. Instead of going big , we explored alternative ways enabled by our immense additional speed-up. For instance, we introduced the concept of <protein mutation movies= (PMMs, Fig. 3) allowing interactively to explore the predicted effects of point mutations (single amino acid variants: SAVs) upon 3D structure (Fig. 2).

The PMMs can benefit from EMBER3D’s speed only because, although less accurate than AlphaFold2, it better predicts effects of sequence variation (Fig. 2). Possibly because AlphaFold2, RoseTTAFold and similar solutions, that began with the PHD first successfully combining machine learning and MSAs 2, learn family-averages rather than protein-specific predictions potentially rendering results less sensitive to small changes in the input (Fig. 2). Also using pLM embeddings, ESMFold clearly surpasses EMBER3D’s accuracy for native sequences. However, ESMFold performed significantly worse in predicting the effect of point mutations (Fig. 2). This might result from the strong regularization of EMBER3D, i.e., its relatively tiny model with 4.7M trainable parameters (Table S1). As a result, small changes in the input are maintained throughout the network without getting mapped to the most similar structural motif seen during training. Even if ESMFold were more sensitive to SAV effects (Fig. 2), it still would not be fast enough to explore a large space of mutants (Fig. 1A), especially when considering to explore higher-order effects such as SAV duplets, triplets, asf.

As proof-of-principle, we pursued several examples of proteins by iteratively picking the SAV that changed the predicted structure most. This naïve heuristic already showed promising results but exploring different sampling strategies to navigate the energy landscape of mutants will be an important future research direction. In this context, the question will not really be how much speedup over AlphaFold2/ColabFold suffices for PMMs , but which aspect becomes crucial first: either the lack of resources to optimize another N-th iteration (with N=3 for triplets) or the limited accuracy in predicting structure (errors might somehow cancel out for N>1 or might increase by error to the Nth power).

Speed may not trump accuracy. ESMFold speeds up over AlpaFold2 at some costs to performance, EMBER3D speeds up more at more costs; random predictions would be instantaneous at higher loss. At which loss of performance is which gain in speed still a trump? There is no proxy for such a point, instead the answer depends on what the method is used for. PMMs are just one aspect. Another aspect is the usage of predicted structure to unravel unsuspected relations between protein pairs through the help of methods such as Foldseek 40 (Fig. 4). In order to leverage the improved sensitive of structure- over sequence-based search methods on a large scale, speedy structure predictions will be crucial as one needs to keep pace with the increasingly growing sequence databases. Another consideration could be the costs to the environment, e.g., by calculating the costs in terms of energy consumption penciling in the hardware and hardware-maintenance costs. Unfortunately, this is not well-defined because the speed advantage of pLM-based solutions shrinks non-linearly with protein length. Consequently, the gain will be much higher for predicting five bacterial than for the human proteome although both may have similar numbers of proteins, and they will be much higher for kinases than for structural proteins.

What makes EMBER3D faster? Presumably, the most obvious factor for EMBER3D outpacing others was the model with relatively few free parameters (Table S3). However, the total parameter count might suggest our model to be relatively slow, given its size (1.5B parameters). However, the underlying pLM, ProtT5 11, accounts for 99.7% of those parameters, leaving only 0.3% (4.7M) for the actual structure prediction module. The frozen pLM (no gradients are backpropagated to the pLM, i.e., it is only used as static feature encoder), however, makes up only a small fraction of the compute time, especially, as it is run in half-precision, turning the structure module into the bottleneck for speed. Therefore, we kept this part of the model as small as possible. We further benefited from an optimized version of the SE( 3 ) module provided by Nvidia that allowed to roughly half inference time at reduced memory consumption. We also reduced the computational complexity of the structure prediction module by only modeling backbone atoms. While neglecting side-chains loses information, for some use-cases the resulting speed-up compensated for this loss. Better for proteins from small families? Although EMBER3D is much less accurate than AlphaFold2 (Fig. 1) and ESMFold 21 and RGN2 22 are a little less accurate 21, all three appear to outperform MSAbased methods when tested on single sequence (Fig. 1). However, strictly speaking, at least for ESMFold and EMBER3D, this claim was a proxy: we simply tested what would happen if we took the proteins for which we do have large MSAs and pretended that we do not have those. We had to refrain to this proxy because there are no high resolution structures available for proteins without MSAs because experimental structural biologists successfully optimize the leverage of each structure 44, i.e., annotations exist more likely for large families. Larger families may be easier to predict than smaller ones. In fact, this may exactly explain why proteins from the largest families tend to be predicted better by pLM- than by MSA-based methods 5,11 which clearly contain more information 45. Thus, our proxy (Fig. 1: <no MSA=) may not capture the full story. Nevertheless, the fraction of proteins most interesting for the novelty-value they carry is likely to be substantially higher for understudied pathogens and species than for the human proteome. In other words, for many proteins that might become relevant for health and biotechnology, predictions not using MSAs may provide much better answers than the state-of-the-art based on evolutionary information. Do mutation movies capture any aspect of structural plasticity? One important direction of protein research is the <hallucination of new structures=46. Could the mutation movies we introduced here help to trace evolution in the tube (e.g., the evolution from one shape to another through sequence variation), or to find new solutions fitting to particular constraints? More basically: do differences between mutant and wild type capture any aspects of the free energy landscape of proteins47? If so, would those aspects matter for understanding the problems of protein dynamics theoretically and/or would they have any practical impact? The correlation between functional DMS data and differences in structures predicted for mutants appeared encouraging, in particular, given that the DMS assays function, i.e., even experimentally measured structural differences may not result in higher levels of correlation.

EMBER3D captured effects of sequence variation upon structure better than methods such as AlphaFold2 although it predicts structures for wild types with sufficient evolutionary information better. This clearly implied one advance of our new technology toward understanding protein function.

Conclusions

We introduced a novel structure prediction system, EMBER3D , computing both 2D (distance maps) and 3D structure (backbone coordinates) from sequence alone in milliseconds for average-length proteins using consumer-grade hardware, based on embeddings from protein Language Models (pLMs), in particular on ProtT5. While the accuracy of EMBER3D does not rival state-of-the-art systems with focus on quality, such as AlphaFold2, RoseTTAFold, or even ESMFold, our method is more sensitive to small changes in amino-acid sequence and therefore useful as a predictor for structural effects of point mutations (single amino acid variants, SAVs) or combinations thereof. 12 This ability, paired with its speed, allows EMBER3D to render almost-real-time protein mutation movies (PMM) that visualize the predicted structural effect of each possible SAV in a protein. Effectively, this enables live-editing of sequences with near-instantaneous feedback of the impact on the predicted structure, e.g., in a webserver. Despite its lower accuracy, EMBER3D predictions suffice for highly sensitive structure search through Foldseek , i.e., successful root the identification of proteins of similar structure through prediction. Many proteins relevant for novel discoveries might be in small families for which EMBER3D appears to predict not only faster but also better than AlphaFold2. EMBER3D’s lightning-fast structure predictions open the door to a variety of new use-cases and possibly could lead to finding mutation paths that switch proteins between different structural conformations.

Methods

Dataset. We obtained 102,823 sequences from SidechainNet 48 as the base for training and validation; SidechainNet adds torsion angles and coordinates for all heavy atoms to ProteinNet12 49. Due memory constraints, we removed proteins >430 residues, yielding 93,286 sequences for training. To avoid bias from overrepresented families, we clustered the training set with UniqueProt 50 at the default Hval=0. Following AlphaFold developers 4, we trained on randomly picked samples from the resulting 11,580 clusters. We optimized hyperparameters on the in-built validation set of SidechainNet with 39 CASP12 51 proteins. We assessed the performance of the final model on 18 CASP1452 domains from free-modeling (FM) and template-based modeling hard (TBM-hard). ProtT5 protein language model . In this work, we generated embeddings for each protein sequence using the pLM ProtT5-XL-UniRef50 11 (for simplicity ProtT5) built in analogy to the NLP model T553. ProtT5 was exclusively trained on unlabelled protein sequences from BFD (Big Fantastic Database; 2.5 billion sequences including meta-genomic sequences) 54 and UniRef50 55. Ultimately, this allowed ProtT5 to learn constraints imprinted by structure and function upon protein sequences. As ProtT5 used only unlabelled sequences and we added no supervised training or finetuning, there was no risk of information leakage or overfitting to the tasks addressed by EMBER. Model architecture . We adopted a similar approach as RoseTTAFold 42 by using 2-track blocks, processing 1D and 2D features, followed by a series of 3-track structure blocks, jointly processing 1D, 2D and 3D information (Fig. S1). In contrast to RoseTTAFold, we used only one 2-track and two 3-track blocks to minimize runtime and memory.

We also replaced the original SE( 3 ) transformer from RoseTTAFold by a more efficient NVIDIA implementation 56. Reducing the number of feature channels in the 1D and 2D pipeline reduced memory further; this allowed processing longer proteins on GPUs with less vRAM. Besides those changes, the internal layout of the attention blocks in the 2- and 3-track blocks of EMBER3D was identical to RoseTTAFold.

Input, output, training . While RoseTTAFold inputs MSAs and template information to the 1D and 2D parts of the network, we solely relied on features derived by ProtT5 from single protein sequences. We extracted 1D representations in the form of ProtT5 embeddings and 2D representations from the ProtT5 attention heads as described in more detail elsewhere 5.

EMBER3D predicts C-beta distograms (42 bins), anglegrams (omega & theta: 37 bins, phi: 19 bins), and 3D coordinates in PDB format for all backbone atoms (C, C-alpha, N, O). As others 4,42, we estimated confidence by predicting the lDDT score for each residue, and stored predicted lDDT scores in the b-value column of the output PDB file.

EMBER was trained with Adamax at a learning rate of 0.001 until the lDDT score of the 3D predictions on the CASP12 validation set showed no improvement over 10 epochs. Relaxing predicted structures . In addition to the backbone coordinates obtained directly from our network, we used pyRosetta 57 to relax 3D models based on the predicted C-beta distance maps and angle maps. We first generated 75 different decoys using short-, mid- and long-range distances from the predicted distograms at varying levels of distance probability thresholds (here: [0.05, 0.45]). We then computed quality estimates of lDDT scores using DeepAccNet 45 to select the best predicted fold.

Data availability

Pre-trained models, inference code, the mutation movie tool, a demo webserver and further resources are publicly available at https://github.com/kWeissenow/EMBER3D. A Google Colab notebook for structure prediction and mutation movie generation can be found at https://colab.research.google.com/drive/16qMVCRKPSLPI08vLxVZnBEB70qYKLqTV.

Acknowledgements

We thank Tim Karl (TUM) for invaluable help with hard- and software and Inga Weise (TUM) for supporting many aspects of this work. Thanks to Minkyung Baek and the co-developers from Baker Lab for publishing the RoseTTAFold source code; thanks to Milo Mirdita (Seoul) for his contribution to making AlphaFold2 available through ColabFold. We gratefully acknowledge the support of NVIDIA with the donation of a Titan GPU used for development. Last not least, thanks to all who make their experimental data publicly available and all those who maintain such databases, in particular to Steve Burley and his team at the PDB. 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 3 4 5 6 7 8 9 10 11 12 13 14 15 4 5 6 7 8 9 10 11 12 13 14 15 16 18 19 20 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37

1 Moult, J. , Pedersen , J. T. , Judson , R. & Fidelis , K. A large-scale experiment to assess protein structure prediction methods . Proteins 23 , ii-v ( 1995 ). https://doi.org: 10 .1002/prot.340230303 2 Rost, B. & Sander , C. Prediction of protein secondary structure at better than 70% accuracy . Journal of Molecular Biology 232 , 584 - 599 ( 1993 ). 3 Marx, V. Method of the Year: protein structure prediction . Nat Methods 19 , 5 - 10 ( 2022 ). https://doi.org: 10 .1038/s41592-021-01359-1 4 Jumper, J. et al. Highly accurate protein structure prediction with AlphaFold . Nature ( 2021 ). https://doi.org: 10 .1038/s41586-021-03819-2 5 Weissenow, K. , Heinzinger , M. & Rost , B. Protein language -model embeddings for fast, accurate, and alignment-free protein structure prediction . Structure 30 , 1169 - 1177 . e1164 ( 2022 ). https://doi.org: 10 .1016/j.str. 2022 . 05 .001 6 Alley, E. C. , Khimulya , G. , Biswas , S. , AlQuraishi, M. & Church , G. M. Unified rational protein engineering with sequence-based deep representation learning . Nat Methods 16 , 1315 - 1322 ( 2019 ). https://doi.org: 10 .1038/s41592-019-0598-1 7 Bepler, T. & Berger , B. Learning protein sequence embeddings using information from structure . arXiv ( 2019 ). https://doi.org:arXiv: 1902 .08661 8 Heinzinger, M. et al. Modeling aspects of the language of life through transfer-learning protein sequences . BMC Bioinformatics 20 , 723 ( 2019 ). https://doi.org: 10 .1186/s12859-019-3220-8 9 Bepler, T. & Berger , B. Learning the protein language: Evolution, structure, and function . Cell Syst 12 , 654 - 669 e653 ( 2021 ). https://doi.org: 10 .1016/j.cels. 2021 . 05 .017 10 Rives, A. et al. Biological structure and function emerge from scaling unsupervised learning to 250 million protein sequences . Proc Natl Acad Sci U S A 118 ( 2021 ). https://doi.org: 10 .1073/pnas.2016239118 11 Elnaggar, A. et al. ProtTrans: Towards Cracking the Language of Lifes Code Through Self-Supervised Deep Learning and High Performance Computing . IEEE Trans Pattern Anal Mach Intell ( 2021 ). https://doi.org: 10 .1109/TPAMI. 2021 .3095381 12 Rao, R. , Meier , J. , Sercu , T. , Ovchinnikov , S. & Rives , A. Transformer protein language models are unsupervised structure learners . bioRxiv , 2020 . 2012 . 2015 . 422761 ( 2020 ). https://doi.org: 10 .1101/ 2020 .12.15.422761 13 Ilzhoefer, D. , Heinzinger , M. & Rost , B. (bioRxiv, 2022 ). 14 Littmann, M. , Heinzinger , M. , Dallago , C. , Weissenow , K. & Rost , B. Protein embeddings and deep learning predict binding residues for various ligand classes . Scientific Reports 11 , 23916 ( 2021 ). https://doi.org: 10 .1038/s41598-021-03431-4 15 Littmann, M. , Heinzinger , M. , Dallago , C. , Olenyi , T. & Rost , B. Embeddings from deep learning transfer GO annotations beyond homology . Scientific Reports 11 , 1160 ( 2021 ). https://doi.org: 10 .1038/s41598- 020-80786-0 16 Littmann, M. et al. Clustering FunFams using sequence embeddings improves EC purity Bioinformatics ( 2021 ). https://doi.org:https://doi.org/10.1093/bioinformatics/btab371 17 Heinzinger, M. et al. Contrastive learning on protein embeddings enlightens midnight zone . NAR Genomics and Bioinformatics 4 , lqac043 ( 2022 ). https://doi.org:https://doi.org/10.1093/nargab/lqac043 18 Bileschi, M. L. et al. Using deep learning to annotate the protein universe . Nature Biotechnology 40 , 932 - 937 ( 2022 ). https://doi.org: 10 .1038/s41587-021-01179-w 19 Wang, G. a. F. , Xiaomin and Wu, Zhihua and Liu, Yiqun and Xue, Yang and Xiang, Yingfei and Yu, Dianhai and Wang, Fan and Ma, Yanjun. HelixFold: An Efficient Implementation of AlphaFold2 using PaddlePaddle . ( 2022 ). https://doi.org: 10 .48550/ARXIV.2207.05477 20 Wu, R. et al. High-resolution <em>de novo</em> structure prediction from primary sequence . bioRxiv , 2022 . 2007 . 2021 . 500999 ( 2022 ). https://doi.org: 10 .1101/ 2022 .07.21.500999 21 Lin, Z. et al. Language models of protein sequences at the scale of evolution enable accurate structure prediction . bioRxiv , 2022 . 2007 . 2020 . 500902 ( 2022 ). https://doi.org: 10 .1101/ 2022 .07.20.500902 22 Chowdhury, R. et al. Single-sequence protein structure prediction using a language model and deep learning . Nat Biotechnol 40 , 1617 - 1623 ( 2022 ). https://doi.org: 10 .1038/s41587-022-01432-w 23 Tunyasuvunakool, K. et al. Highly accurate protein structure prediction for the human proteome . Nature ( 2021 ). https://doi.org: 10 .1038/s41586-021-03828-1 24 Mirdita, M. et al. ColabFold - Making protein folding accessible to all . bioRxiv , 2021 . 2008 . 2015 . 456425 ( 2021 ). https://doi.org: 10 .1101/ 2021 .08.15.456425 25 Kryshtafovych, A. , Schwede , T. , Topf , M. , Fidelis , K. & Moult , J. Critical assessment of methods of protein structure prediction (CASP)4Round XIII. Proteins: Structure , Function, and Bioinformatics 87 , 1011 - 1020 ( 2019 ). https://doi.org:https://doi.org/10.1002/prot.25823 26 Consortium, T. U. UniProt: the universal protein knowledgebase in 2021 . Nucleic Acids Research 49 , D480 - D489 ( 2020 ). https://doi.org: 10 .1093/nar/gkaa1100 27 Mariani, V. , Biasini , M. , Barbato , A. & Schwede , T. lDDT: a local superposition-free score for comparing protein structures and models using distance difference tests . Bioinformatics 29 , 2722 - 2728 ( 2013 ). https://doi.org: 10 .1093/bioinformatics/btt473 28 Fowler, D. M. & Fields , S. Deep mutational scanning: a new style of protein science . Nature Methods 11 , 801 - 807 ( 2014 ). https://doi.org: 10 .1038/nmeth.3027 29 Bandaru, P. et al. Deconstruction of the Ras switching cycle through saturation mutagenesis . eLife 6 , e27810 ( 2017 ). https://doi.org: 10 .7554/eLife.27810 30 Mavor, D. et al. Determination of ubiquitin fitness landscapes under different chemical stresses in a classroom setting . eLife 5 , e15802 ( 2016 ). https://doi.org: 10 .7554/eLife.15802 31 Kelsic, E. D. et al. RNA Structural Determinants of Optimal Codons Revealed by MAGE-Seq . Cell Systems 3 , 563 - 571 . e566 ( 2016 ). https://doi.org:https://doi.org/10.1016/j.cels. 2016 . 11 .004 32 Weile, J. et al. A framework for exhaustively mapping functional missense variants . Molecular Systems Biology 13 , 957 ( 2017 ). https://doi.org:https://doi.org/10.15252/msb.20177908 33 Chan, Y. H. , Venev , S. V. , Zeldovich , K. B. & Matthews , C. R. Correlation of fitness landscapes from three orthologous TIM barrels originates from sequence and structure constraints . Nature Communications 8 , 14614 ( 2017 ). https://doi.org: 10 .1038/ncomms14614 34 Suiter, C. C. et al. Massively parallel variant characterization identifies <i>NUDT15</i> alleles associated with thiopurine toxicity . Proceedings of the National Academy of Sciences 117 , 5394 - 5401 ( 2020 ). https://doi.org:doi:10.1073/pnas.1915680117 35 Hunter, J. D. Matplotlib : A 2D Graphics Environment . Computing in Science & Engineering 9 , 90 - 95 ( 2007 ). https://doi.org: 10 .1109/ MCSE . 2007 .55 36 Schrödinger, L. & DeLano , W. The PyMOL Molecular Graphics System , <http://www.pymol.org/pymol> ( 2021 ). 37 Tomar, S. Converting video formats with FFmpeg . Linux Journal 2006 , 10 ( 2006 ). 38 Williams, S. G. & Lovell , S. C. The Effect of Sequence Evolution on Protein Structural Divergence . Molecular Biology and Evolution 26 , 1055 - 1065 ( 2009 ). https://doi.org: 10 .1093/molbev/msp020 39 Lo Conte , L. et al. SCOP: a structural classification of proteins database . Nucleic Acids Res 28 , 257 - 259 ( 2000 ). https://doi.org: 10 .1093/nar/28.1. 257 40 van Kempen , M. et al. Foldseek: fast and accurate protein structure search . bioRxiv , 2022 . 2002 . 2007 . 479398 ( 2022 ). https://doi.org: 10 .1101/ 2022 .02.07.479398 41 Steinegger, M. & Söding , J. MMseqs2 enables sensitive protein sequence searching for the analysis of massive data sets . Nature Biotechnology 35 , 1026 - 1028 ( 2017 ). https://doi.org: 10 .1038/nbt.3988 42 Baek, M. et al. Accurate prediction of protein structures and interactions using a three-track neural network . Science 373 , 871 - 876 ( 2021 ). https://doi.org:doi:10.1126/science.abj8754 43 Callaway, E. AlphaFold's new rival? Meta AI predicts shape of 600 million proteins . Nature 611 , 211 - 212 ( 2022 ). 44 Liu, J. , Montelione , G. T. & Rost , B. Novel leverage of structural genomics . Nature Biotechnology 25 , 849 - 851 ( 2007 ). 45 Hiranuma, N. et al. Improved protein structure refinement guided by deep learning based accuracy estimation . Nature Communications 12 , 1340 ( 2021 ). https://doi.org: 10 .1038/s41467-021-21511-x 46 Anishchenko, I. et al. De novo protein design by deep network hallucination . Nature 600 , 547 - 552 ( 2021 ). https://doi.org: 10 .1038/s41586-021-04184-w 47 Onuchic, J. N. , Luthey-Schulten , Z. & Wolynes , P. G. Theory of protein folding: the energy landscape perspective . Annu Rev Phys Chem 48 , 545 - 600 ( 1997 ). https://doi.org: 10 .1146/annurev.physchem. 48 .1. 545 48 King, J. E. & Koes , D. R. SidechainNet: An all-atom protein structure dataset for machine learning . Proteins: Structure, Function, and Bioinformatics 89 , 1489 - 1496 ( 2021 ). https://doi.org:https://doi.org/10.1002/prot.26169 49 AlQuraishi, M. ProteinNet: a standardized data set for machine learning of protein structure . BMC Bioinformatics 20 , 311 ( 2019 ). https://doi.org: 10 .1186/s12859-019-2932-0 50 Olenyi, T. a. B. , Michael and Mirdita, Milot and Steinegger, Martin and Rost, Burkhard. Rostclust - - Protein Redundancy Reduction (School of Computation, Information, and Technology , Technical University of Munich., 2022 ). 51 Moult, J. , Fidelis , K. , Kryshtafovych , A. , Schwede , T. & Tramontano , A. Critical assessment of methods of protein structure prediction (CASP)-Round XII . Proteins 86 Suppl 1 , 7 - 15 ( 2018 ). https://doi.org: 10 .1002/prot.25415 52 Pereira, J. et al. High-accuracy protein structure prediction in CASP14 . Proteins 89 , 1687 - 1699 ( 2021 ). https://doi.org: 10 .1002/prot.26171 53 Raffel, C. et al. Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer . arXiv ( 2020 ). 54 Steinegger, M. , Mirdita , M. & Söding , J. Protein-level assembly increases protein sequence recovery from metagenomic samples manyfold . Nature Methods 16 , 603 - 606 ( 2019 ). https://doi.org: 10 .1038/s41592- 019-0437-4 55 Suzek, B. E. , Wang , Y. , Huang , H. , McGarvey , P. B. & Wu , C. H. UniRef clusters: a comprehensive and scalable alternative for improving sequence similarity searches . Bioinformatics 31 , 926 - 932 ( 2015 ). https://doi.org: 10 .1093/bioinformatics/btu739 56 Milesi, A. Accelerating SE ( 3)-Transformers Training Using an NVIDIA Open-Source Model Implementation . ( 2021 ). <https://developer.nvidia.com/blog/accelerating-se3 - transformers -trainingusing-an-nvidia-open-source-model-implementation/>. 57 Chaudhury, S. , Lyskov , S. & Gray , J. J. PyRosetta: a script-based interface for implementing molecular modeling algorithms using Rosetta . Bioinformatics 26 , 689 - 691 ( 2010 ). https://doi.org: 10 .1093/bioinformatics/btq007 58 Zhang, Y. & Skolnick , J. TM-align: a protein structure alignment algorithm based on the TM-score . Nucleic Acids Research 33 , 2302 - 2309 ( 2005 ). https://doi.org: 10 .1093/nar/gki524 59 Harris, C. R. et al. Array programming with NumPy . Nature 585 , 357 - 362 ( 2020 ). https://doi.org: 10 .1038/s41586-020-2649-2