October ProteInfer: deep networks for protein functional inference Theo Sanderson 0 1 Maxwell L. Bileschi 0 David Belanger Lucy J. Colwell Equal contribution Francis Crick Institute 2021 6 2021

Predicting the function of a protein from its amino acid sequence is a long-standing challenge in bioinformatics. Traditional approaches use sequence alignment to compare a query sequence either to thousands of models of protein families or to large databases of individual protein sequences. Here we instead employ deep convolutional neural networks to directly predict a variety of protein functions - EC numbers and GO terms - directly from an unaligned amino acid sequence. This approach provides precise predictions which complement alignment-based methods, and the computational efficiency of a single neural network permits novel and lightweight software interfaces, which we demonstrate with an in-browser graphical interface for protein function prediction in which all computation is performed on the user's personal computer with no data uploaded to remote servers. Moreover, these models place full-length amino acid sequences into a generalised functional space, facilitating downstream analysis and interpretation. To read the interactive version of this paper, please visit https: //google-research.github.io/proteinfer/ Sanderson et al. | bioRχiv | October 6, 2021 | 1-21

-

QR code for the interactive version of this preprint at https: //google-research.github.io/proteinfer/

Introduction

Every day, more than a hundred thousand protein sequences are added to global sequence databases ( 1 ). However, these entries are of limited use to practitioners unless they are accompanied by functional annotations. While curators diligently extract annotations from the literature, assessing more than 60,000 papers each year ( 2 ), the time-consuming nature of this task means that only 0.03% of publicly available protein sequences are manually annotated. The community has a long history of using computational tools to infer protein function directly from amino acid sequence. Starting in the 1980s, methods such as BLAST ( 3 ) relied on pairwise sequence comparisons, where a query protein is assumed to

BLAST Find closest seq in large DB, transfer annotations

MEGERSTCFYY FWAWLPINAWE LPMEEEFSTPI LNRRFLAWAFL Input: amino acid sequence

InterProScan

Apply >50,000 signatures & HMMs - Amylase activity - Exported protein Output: annotation with a subset of >10,000 multi-hot labels

Apply single

deep neural network

ProteInfer

Fig. 1. Three approaches for mapping from an amino acid sequence to inferred function: ( 1 ) finding similar sequences in a large database of sequences with known annotation (e.g., BLAST), ( 2 ) scoring against a large database of statistical models for each family of sequences with known function (e.g. InterProScan), and ( 3 ) applying a single deep neural network trained to predict multiple output categories (e.g., this work). have the same function as highly similar sequences that have already been annotated. Signature-based approaches were later introduced, with the PROSITE database ( 4 ) cataloguing short amino acid “motifs” found in proteins that share a particular function. Subsequently, a crucial refinement of signature-based approaches was the development of profile hidden Markov models (HMMs) ( 5, 6 ). These models collapse an alignment of related protein sequences into a model that provides likelihood scores for new sequences, that describe how well they fit the aligned set. Critically, profile HMMs allow for longer signatures and fuzzier matching and are currently used to update popular databases such as Interpro and Pfam ( 7, 8 ). Subsequent refinements have made these techniques more sensitive and computationally efficient ( 7, 9–12 ), while their availability as web tools allows practitioners to easily incorporate them into workflows ( 13–16 ). These computational modelling approaches have had great impact; however, one third of bacterial proteins still cannot be annotated (even computationally) with a function ( 17 ). It is therefore worthwhile to examine how new approaches might complement existing techniques. First, current approaches conduct entirely separate comparisons for each comparator sequence or model, and thus may not fully exploit the features shared across different functional classes. An ideal classification system, for example, might have a modular ATPbinding region detector used in detection of both kinases and ABC transporters ( 18 ). Separately modelling these targets in each family, like standard HMM approaches, increases the computational cost and may also be less accurate. In addition, the process of creating many of these signatures is not fully automated and requires considerable curatorial efforts ( 19, 20 ), which at present are spread across an array of disparate but overlapping signature databases ( 8 ).

Deep neural networks have recently transformed a number of labelling tasks, including image recognition – the early layers in these models build up an understanding of simple features such as edges, and later layers use these features to identify textures, and then entire objects. Edge detecting filters can thus be trained with information from all the labelled examples, and the same filters can be used to detect, for instance, both oranges and lemons ( 21 ).

In response, recent work has contributed a number of deep neural network models for protein function classification ( 22–30 ). These approaches train a single model to recognise multiple properties, building representations of different protein functions via a series of layers, which allow the same low-level features to be used for different high-level classifications. Of special note is the layer preceding the final layer of the network, which constructs an “embedding” of the entire example in a high-dimensional vector space, and often captures semantic features of the input.

Beyond functional annotation, deep learning has enabled significant advances in protein structure prediction ( 31–36 ), predicting the functional effects of mutations ( 37–40 ), and protein design ( 41–47 ). A key departure from traditional approaches is that researchers have started to incorporate vast amounts of raw, uncurated sequence data into model training, an approach which also shows promise for functional prediction ( 48 ).

Of particular relevance to the present work is Bileschi et al. (2019) ( 49 ), where it is shown that models with residual layers ( 50 ) of dilated convolutions ( 51 ) can precisely and efficiently categorise protein domains. Dohan (2021) ( 52 ) provides additional accuracy improvements using uncurated data. However, these models cannot infer functional annotations for full-length protein sequences, since they are trained on pre-segmented domains and can only predict a single label. The full-sequence task is of primary importance to biological practitioners.

To address this challenge we employ deep dilated convolutional networks to learn the mapping between full-length protein sequences and functional annotations. The resulting ProteInfer models take amino acid sequences as input and are trained on the well-curated portion of the protein universe annotated by Swiss-Prot ( 2 ). We find that: 1) ProteInfer models reproduce curator decisions for a variety of functional properties across sequences distant from the training data, 2) attribution analysis shows that the predictions are driven by relevant regions of each protein sequence, and 3) ProteInfer models create a generalised mapping between sequence space and the space of protein functions, which is useful for tasks other than those for which the models were trained. We provide trained ProteInfer networks that enable other researchers to reproduce the analysis presented and to explore embeddings of their proteins of interest, via both a command line tool1, and also via an in-browser JavaScript implementation that demonstrates the computational efficiency of deep-learning approaches.

Methods

Class outputs EC:1.1.1.1

EC:1.1.1.2

EC:1.1.1.3

EC:1.1.1.4

EC:1.1.1.5

...

mean pool Embedding

n residual convolutional

blocks One-hot

Input M

E G E R S T

C F

Y

A neural-network for protein function prediction In a ProteInfer neural network (Fig. 2), a raw amino acid sequence is first represented numerically as a one-hot matrix and then passed through a series of convolutional layers. Each layer takes the representation of the sequence in the previous layer and applies a number of filters, which detect patterns of features. We use residual layers, in which the output of each layer is added to its input to ease the training of deeper networks ( 50 ), and dilated convolutions ( 51 ), meaning that successive layers examine larger sub-sequences of the input sequence. After building up an embedding of each position in the sequence, the model collapses these down to a single n-dimensional embedding of the sequence using average pooling. Since natural protein sequences can vary in length by at least three orders of magnitude, this pooling is advantageous because it allows our model to accommodate sequences of arbitrary length without imposing restrictive modeling assumptions or computational burdens that scale with sequence length. Finally, a fully-connected layer maps these embeddings to logits for each potential label, which are the input to an element-wise sigmoid layer that outputs perlabel probabilities. We select all labels with predicted probability above a given confidence threshold, and varying this threshold yields a tradeoff between precision and recall. To summarize model performance as a single scalar, we compute the Fmax score, the maximum F1 score (the geometric mean of precision and recall) across all thresholds ( 53 ). Each model was trained for about 60 hours using the Adam optimizer ( 54 ) on 8 NVIDIA P100 GPUs with data parallelism ( 55, 56 ). We found that using more than one GPU 1https://github.com/google-research/proteinfer for training improved training time by allowing an increased batch size, but did not have a substantial impact on accuracy compared to training for longer with a smaller learning rate and smaller batch size on one GPU. The models have a small set of hyperparameters, such as the number of layers and the number of filters in each layer, which were tuned using random sampling to maximize Fmax on the random train-test split. Hyperparameter values are available in the Supplement.

A machine-learning compatible dataset for protein function prediction The UniProt database is the central global repository for information about proteins. The manually curated portion, Swiss-Prot, is constructed by assessing 60,000 papers each year to harvest 35% of the theoretically curatable information in the literature ( 2 ). We focus on Swiss-Prot to ensure that our models learn from human-curated labels, rather than labels generated by a computational annotation pipeline. Each protein in Swiss-Prot goes through a 6-stage process of sequence curation, sequence analysis, literature curation, family-based curation, evidence attribution, and quality assurance. Functional annotation is stored in UniProt largely through database cross-references, which link a specific protein with a label from a particular ontology. These crossreferences include: Enzyme Commission (EC) numbers, representing the function of an enzyme; Gene Ontology (GO) terms relating to the protein’s molecular function, biological process, or subcellular localisation; protein family information contained in the Pfam ( 7 ), SUPFAM( 57 ), PRINTS ( 58 ), TIGR ( 59 ), PANTHR ( 60 ) databases, or the umbrella database InterPro ( 61 ); as well as other information including ortholog databases and references on PubMed. Here, we focus on EC and GO labels, though our model training framework can immediately extend to other label sets.

We use two methods to split data into training and evaluation sets. First, a random split of the data allows us to answer the following question: suppose that curators had randomly annotated only 80% of the sequences in Swiss-Prot. How accurately can ProteInfer annotate the remaining 20%? Second, we use UniRef50 ( 62 ) clustering to split the data, to model a challenging use-case in which an unseen sequence has low sequence similarity to anything that has been previously annotated. Note there are alternative methods for splitting ( 48, 63, 64 ), such as reserving the most recentlyannotated proteins for evaluating models. This approach, which is used in CAFA and CASP ( 63, 64 ), helps ensure a fair competition because labels for the evaluation data are not available to participants, or the scientific community at large, until after the competition submissions are due. Such a split is not available for EC classification, which is the primary focus of our analyses below. Finally, note that all of the above approaches typically lack reliable annotation for true negatives ( 65 ).

To facilitate further development of machine learning methods, we provide TensorFlow ( 66 ) TFrecord files for Swiss

Prot 2. Each example has three fields: the UniProt accession, the amino acid sequence, and a list of database crossreference labels. UniProt annotations include only leaf nodes for hierarchical taxononomies such as EC and GO. To allow machine-learning algorithms to model this hierarchy, we added all parental annotations to each leaf node during dataset creation.

Results

Prediction of catalysed reactions We initially trained a model to predict enzymatic catalytic activities from amino acid sequence. This data is recorded as Enzyme Commission (EC) numbers, which describe a hierarchy of catalytic functions. For instance, β amylase enzymes have an EC number of EC:3.2.1.2, which represents the leaf node in the following hierarchy:

EC:3.–.–.– (hydrolases)

EC:3.2.–.– (glycolsylases)

EC:3.2.1.– (glycosidases)

EC:3.2.1.2 (hydrolysis of

(1→4)-α-Dglucosidic linkages in polysaccharides) Individual protein sequences can be annotated with zero (non-enzymatic proteins), one (enzymes with a single function) or many (multi-functional enzymes) leaf-level EC numbers. These are drawn from a total of 8,162 catalogued chemical reactions. Our best Fmax was achieved by a model containing 5 residual blocks with 1100 filters each (full details in Supplement). For the dev set, Fmax converged within 500,000 training steps. On the random split, the model achieves Fmax = 0.977 (0.976-0.978) on the held-out test data. At the corresponding confidence threshold, the model correctly predicts 96.7% of true labels, with a false positive rate of 1.4%. Results from the clustered test set are discussed below. Performance was roughly similar across labels at the top of the EC hierarchy, with the highest Fmax score observed for ligases (0.993), and the lowest for oxidoreductases

2https://console.cloud.google.com/storage/ browser/brain-genomics-public/research/proteins/ proteinfer/datasets/ (0.963) (Fig. 3A). For all classes, the precision of the network was higher than the recall at the threshold maximising Fmax.

Precision and recall can be traded off against each other by adjusting the confidence threshold at which the network outputs a prediction, creating the curves shown in Fig. 3B.

BLASTp is arguably a practitioner’s default choice for functional annotation, so we implemented an alignment-based baseline in which BLASTp is used to identify the closest sequence to a query sequence in the train set. Labels are then imputed for the query sequence by transferring those labels that apply to the annotated match from the train set.

We produced a precision-recall curve by using the bit score of the closest sequence as a measure of confidence, varying the cutoff above which we retain the imputed labels ( 63, 67 ). We also considered an ensemble of neural networks ( 49 ), where the average of the ensemble elements’ predicted probabilities is used as a confidence score. (See Fig. S7, Fig. S8.) We found that BLASTp was able to achieve higher recall values than ProteInfer for lower precision values, while ProteInfer was able to provide greater precision than BLASTp at lower recall values. We wondered whether a combination of ProteInfer and BLASTp could synergize the best properties of both approaches. We found that even the simple ensembling strategy of rescaling the BLAST bit score by the averages of the ensembled CNNs’ predicted probabilities gave a Fmax score (0.991, 95% confidence interval [CI]: 0.990– 0.992 ) that exceeded that of BLAST (0.984, 95% CI: 0.983– 0.985) or the ensembled CNN (0.981, 95% CI: 0.980–0.982) alone. On the clustered train-test split based on UniRef50 (see clustered in Fig. 3B), we see a performance drop in all methods: this is expected, as remote homology tasks are designed to challenge methods to generalize farther in sequence space. The Fmax score of a single neural network fell to 0.914 (95% CI: 0.913–0.915, precision: 0.959 recall: 0.875), substantially lower than BLAST (0.950, 95% CI: 0.950–0.951), though again an ensemble of both BLAST and ProteInfer outperformed both (0.979, 95% CI: 0.979–0.980).

We find that neural network methods learn different information about proteins than alignment-based methods, and a combination of the two further improves remote homology detection.

We also examined the relationship between the number of examples of a label in the training dataset and the performance of the model. In an image recognition task, this is an important consideration since one image of, say, a dog, can be utterly different to another. Large numbers of labels are therefore required to learn filters that are able to predict members of a class. In contrast, for sequence data we found that even for labels that occurred less than five times in the training set, 58% of examples in the test set were correctly recalled, while achieving a precision of 88%, for an F1 of 0.7 (Fig. S9). High levels of performance are maintained with few training examples because of the evolutionary relationship between sequences, which means that one ortholog of a gene may be similar in sequence to another. The simple BLAST implementation described above also performs well, and better than a single neural network, likely again exploiting the fact that many sequence have close neighbours in sequence space with similar functions. We again find that ensembling the BLAST and ProteInfer outputs provides performance exceeding that of either technique used alone.

Deep models link sequence regions to function Proteins that use separate domains to carry out more than one enzymatic function are particularly useful in interpreting the behaviour of our model. For example, S. cerevisiae fol1 (accession Q4LB35) catalyses three sequential steps of tetrahydrofolate synthesis, using three different protein domains (Fig. 4A). This protein is in our held-out test set, so no information about its labels was directly provided to the model.

To investigate what sequence regions the neural network is using to make its functional predictions, we used class activation mapping (CAM) ( 68 ) to identify the sub-sequences responsible for the model predictions. We found that separate regions of sequence cause the prediction of each enzymatic activity, and that these regions correspond to the known functions of these regions (Fig. 4B). This demonstrates that

UMAP dimension 1

UMAP dimension 1 2 n o i s n e m i d P A M U 2 n o i s n e m i d P A M U

UMAP dimension 1 UMAP dimension 1 our network identifies relevant elements of a sequence in determining function.

We then assessed the ability of this method to more generally localize function within a sequence, even though the model was not trained with any explicit localization information. We selected all enzymes from Swiss-Prot that had two separate leaf-node EC labels for which our model predicted known EC labels, and these labels were mappable to corresponding Pfam labels. For each of these proteins, we obtained coarse-grained functional localization by using CAM to predict the order of the domains in the sequence and compared to the ground truth Pfam domain ordering(see supplement for details of the method). We found that in 296 of 304 (97%) of the cases, we correctly predicted the ordering, though we note that the set of bifunctional enzymes for which this analysis is applicable is limited in its functional diversity (see supplement). Although we did not find that fine-grained, per-residue functional localization arose from our application of CAM, we found that it reliably provided coarse-grained annotation of domains’ order, as supported by Pfam. This experiment suggests that this is a promising future area for research.

Neural networks learn a general-purpose embedding space for protein function Whereas InterProScan compares each sequence against more than 50,000 individual signatures and BLAST compares against an even larger sequence database, ProteInfer uses a single deep model to extract features from sequences that directly predict protein function. One convenient property of this approach is that in the penultimate layer of the network each protein is expressed as a single point in a highdimensional space. To investigate to what extent this space is useful in examining enzymatic function, we used the ProteInfer EC model trained on the random split to embed each test set protein sequence into a 1100-dimensional vector.

To visualise this space, we selected proteins with a single leaf-level EC number and used UMAP to compress their embeddings into two dimensions ( 69 ).

The resulting representation captures the hierarchical nature of EC classification, with the largest clusters in embedding space corresponding to top level EC groupings (Fig. 5A).

These clusters in turn are further divided into sub-regions on the basis of subsequent levels of the EC hierarchy (Fig. 5B).

Exceptions to this rule generally recapitulate biological properties. For instance, Q8RUD6 is annotated as Arsenate reductase (glutaredoxin) (EC:1.20.4.1) ( 70 ) was not placed with

(a) Nucleotide binding (b) Structural constituent of ribosome (c) Intrinsic component of membrane Fig. 6. A neural network trained on enzyme function learns general protein properties, beyond enzymatic activity. This figure shows EC-trained ProteInfer embeddings for all non-enzymatic sequences in the test set, projected using UMAP. To illustrate the structure contained in these embeddings we highlight genes based on GO labels (on which this network was never trained). other oxidoreductases (EC:1.-.-.-) but rather with Sulfurtransferases (EC:2.8.1.-). Q8RUD6 can however act as a sulfertransferase ( 71 ).

Note that the model is directly trained with labels reflecting the EC hierarchy; the structure in Fig. 5 was not discovered automatically from the data. However, we can also ask whether the embedding captures more general protein characteristics, beyond those on which it was directly supervised.

To investigate this, we took the subset of proteins in SwissProt that are non-enzymes, and so lack any EC annotations.

The network would achieve perfect accuracy on these examples if it e.g. mapped all of them to a single embedding that corresponds to zero predicted probability for every enzymatic label. Do these proteins therefore share the same representation in embedding space? The UMAP projection of these sequences’ embeddings revealed clear structure to the embedding space, which we visualised by highlighting several GO annotations which the network was never supervised on. For example, one region of the embedding space contained ribosomal proteins, while other regions could be identified containing nucleotide binding proteins, or membrane proteins (Fig. 6). To quantitatively measure whether these embeddings capture the function of non-enzyme proteins, we trained a simple random forest classification model that used these embeddings to predict whether a protein was annotated with the intrinsic component of membrane GO term.

We trained on a small set of non-enzymes containing 518 membrane proteins, and evaluated on the rest of the examples. This simple model achieved a precision of 97% and recall of 60% for an F1 score of 0.74. Model training and data-labelling took around 15 seconds. This demonstrates the power of embeddings to simplify other studies with limited labeled data, as has been observed in recent work ( 43, 72 ).

Rapid client-side in-browser protein function prediction Processing speed and ease-of-access are important considerations for the utility of biological software. An algorithm that takes hours or minutes is less useful than one that runs in seconds, both because of its increased computational cost, but also because it allows less immediate interactivity with a researcher. An ideal tool for protein function prediction would require minimal installation and would instantly answer a biologist’s question about protein function, allowing them to immediately act on the basis of this knowledge. Moreover, there may be intellectual property concerns in sending sequence data to remote servers, so a tool that does annotation completely client-side may also be preferable.

There is arguably room for improvement in this regard from classical approaches. For example, the online interface to InterProScan can take 147 seconds to process a 1500 amino acid sequence34, while running the tool may make the search faster, doing so requires downloading a 9 GB file, with an additional 14 GB for the full set of signatures, which when installed exceeds 51 GB. Meanwhile, conducting a BLAST search against Swiss-Prot takes 34 seconds for a 1500 amino acid sequence 5.

An attractive property of deep learning models is that they can be run efficiently, using consumer graphics cards for acceleration. Indeed, recently, a framework has been developed to allow models developed in TensorFlow to be run locally using simply a user’s browser ( 73 ), but to our knowledge this has never been deployed to investigate biological sequence data. We therefore built a tool to allow near-instantaneous prediction of protein functional properties in the browser.

When the user loads the tool, lightweight EC (5MB) and GO model (7MB) prediction models are downloaded and all predictions are then performed locally, with query sequences never leaving the user’s computer. Inference in the browser for a 1500 amino-acid sequence takes < 1.5 seconds for both models (Sup. Note G).

Comparison to experimental data To assess our model’s performance using an additional experimentally-validated source of ground truth, we focused our attention on a large set of bacterial genes for which functions have recently been identified in a high-throughput experimental genetic study ( 17 ). In particular, this study listed newly identified EC numbers for 171 proteins, representing cases when there was previously either misannotation or inconsistent annotation in the SEED or KEGG databases.

3Protein Q77Z83.

4It should be noted that the individual databases that make up InterProScan may return matches faster, with the online interface to Pfam taking 14–20 seconds for a 1500 amino acid sequence.

5A target database with decreased redundancy could be built to reduce this search time, and other optimizations of BLAST have been developed.

Therefore, this set of genes may be enriched for proteins whose functions are difficult to assess computationally. We examined how well our network was able to make predictions for this experimental dataset at each level of the EC hierarchy (Fig. 7), using as a decision threshold the value that optimised F1 identified during tuning. The network had high accuracy for identification of broad enzyme class, with 90% accuracy at the top level of the EC hierarchy. To compute accuracy, we examined the subset of these 171 proteins for which there was a single enzymatic annotation from ( 17 ), giving us predictions for 119 enzymes. At the second level of the hierarchy, accuracy was 90% and the network declined to make a prediction for 12% of classes. Even at the third level, accuracy was 86% with the network making a prediction in 77% of cases. At the finest level of classification, the proportion of examples for which a prediction was made fell to 28%, with 42% of these predictions correct.

As an example, the Sinorhizobium meliloti protein Q92SI0 is annotated in UniProt as a Inosine-uridine nucleoside Nribohydrolase (EC 3.2.2.1). Analysing the gene with InterProScan ( 74 ) also gives this prediction, but our model instead predicts it to be a uridine nucleosidase (EC 3.2.2.3), and this was indeed the result found in this experimental work. Similarly, Pseudomonas fluorescens A0A166Q345 was correctly classified by our model as a D-galacturonate dehydrogenase (EC 1.1.1.203) as opposed to a misannotation on UniProt and with InterProScan.

It was notable that for many of these proteins, the network declined to make a prediction at the finest level of the EC hierarchy. This suggests that by training on this hierarchical data, the network is able to appropriately make broad or narrow classification decisions. This is similar to the procedure employed with manual annotation: when annotators are confident of the general class of reaction that an enzyme catalyses but not its specific substrate, they may leave the third or fourth position of the EC number blank (e.g. EC:1.1.-.-).

Due to training on hierarchical data, our network is able to reproduce these effects by being more confident (with higher accuracy) at earlier levels of classification. A model predicting the entire gene ontology Given the high accuracy that our deep learning model was able to achieve on the more than five thousand enzymatic labels in Swiss-Prot, we asked whether our networks could learn to predict protein properties using an even larger vocabulary of labels, using a similar test-train setup. Gene Ontology ( 75–77 ) (GO) terms describe important protein functional properties, with 28,079 such terms in Swiss-Prot that cover the molecular functions of proteins (e.g. DNA-binding, amylase activity), the biological processes they are involved in (e.g. DNA replication, meiosis), and the cellular components to which they localise (e.g. mitochondrion, cytosol).

These terms are arranged in a complex directed acyclic graph, with some nodes having as many as 12 ancestral nodes.

We note that there has been extensive work in GO label prediction evaluated on a temporally-split dataset (constructing a test set with the most recently experimentally annotated proteins), e.g., ( 63 ), and stress that our comparison is based on the random and clustered splits of Swiss-Prot described above. This approach to splitting the data into train and test has advantages and disadvantages as compared to a temporal split, which depend on the desired application for the method being evaluated.

We trained a single model to predict presence or absence for each of these terms and found that our network was able to achieve a precision of 0.918 and a recall of 0.854 for an F1 score of 0.885 (95% CI: 0.882–0.887).

An ensemble of multiple CNN elements was again able to achieve a slightly better result with an F1 score of 0.899 (95% CI: 0.897–0.901), which was exceeded by a simple transfer of the BLAST top pick at 0.902 (95% CI: 0.900–0.904), with an ensemble of both producing the best result of 0.908 (95% CI: 0.906–0.911).

To benchmark against a common signature-based methodology, we used InterProScan to assign protein family signatures to each test sequence. We chose InterProScan for its coverage of labels as well as its use of multiple profile-based annotation methods, including HMMER and PROSITE, mentioned above. We note that while InterProScan predicts GO labels directly, it does not do so for EC labels, which is why we did not use InterProScan to benchmark our work on predicting EC labels. We found that InterProScan gave good precision, but within this UniProt data had lower recall, giving it a precision of 0.937 and recall of 0.543 for an F1 score of 0.688. ProteInfer’s recall at a precision of .937 is substantially higher (0.835) than InterProScan at assigning GO labels.

There are multiple caveats to these comparisons. One challenge is that the completeness of Swiss-Prot’s GO term annotations varies ( 78 ). As an extreme example, Pan paniscus (Pygmy Chimpanzee) and Pan troglodytes (Chimpanzee) have an identical Apolipoprotein A-II protein,6, where the first protein has 24 GO annotations, while the latter has 143 GO annotations.7 One way this is reflected in the perfor6Accessions P0DM95 and Q8MIQ5.

7This count is done using not only the set of all labels that appear in SwissProt, but also any parents of those labels. mance of the models is that some BLAST matches that have extremely large bit-scores are not annotated identically, and thus reduce the precision of the BLAST model. It is also important to note that our model has the advantage of being specifically trained on the UniProt labelling schema upon which it is being evaluated. InterPro works quite differently, with GO terms being assigned to families, and so inconsistencies in terms of how these are assigned can explain reduced performance – for instance InterPro families simply do not feature all of the GO terms found in UniProt. Thus these results should be seen as specific to the task of reproducing the curated results in UniProt.

Discussion

We have shown that neural networks trained and evaluated on high-quality Swiss-Prot data accurately predict functional properties of proteins using only their raw, un-aligned amino acid sequences. Further, our models make links between the regions of a protein and the function that they confer, produce predictions that agree with experimental characterisations, and place proteins into an embedding space that captures additional properties beyond those on which the models were directly trained. We have provided a convenient browser-based tool, where all computation runs locally on the user’s computer. To support follow-up research, we have also released our datasets, code for model training and evaluation, and a command-line version of the tool.

Using Swiss-Prot to benchmark our tool against traditional alignment-based methods has distinct advantages and disadvantages. It is desirable because the data has been carefully curated by experts, and thus it contains minimal falsepositives. On the other hand, many entries come from experts applying existing computational methods, including BLAST and HMM-based approaches, to identify protein function. Therefore, the data may be enriched for sequences with functions that are easily ascribable using these techniqueswhich could limit the ability to estimate the added value of using an alternative alignment-free tool. An idealised dataset would involved training only on those sequences that have themselves been experimentally characterized, but at present too little data exists than would be needed for a fully supervised deep-learning approach. Semi-supervised approaches that combine a smaller number of high quality experimental labels with the vast set of amino acid sequences in TrEMBL may be a productive way forward.

Further, our work characterizes proteins by assigning labels from a fixed, pre-defined set, but there are many proteins with functions that are not covered by this set. These categories of functions may not even be known to the scientific community yet. There is a large body of alternative work that identifies groups of related sequences (e.g. ( 79 )), where a novel function could be discovered, for example, using follow-up experiments.

Finally, despite the successes of deep learning in many application domains, a number of troublesome behaviours have also been identified. For example, probabilities output by deep models are often over-confident, rather than well-calibrated ( 80 ), and networks perform poorly on outof-distribution data without being aware that they are outside their own range of expertise ( 81 ). Though these issues still need to be addressed and better understood by both the machine learning and bioinformatics communities, deep learning continues to make advances in a wide range of areas relating to the understanding protein function. We thus believe deep learning will have a central place in the future of this field.

Our code, data, and notebooks reproducing the analyses shown in this work are available online at https: //github.com/google-research/proteinfer and https://console.cloud.google.com/ storage/browser/brain-genomics-public/ research/proteins/proteinfer/datasets/. ACKNOWLEDGEMENTS We would like to thank Babak Alipanahi, Jamie Smith, Eli Bixby, Drew Bryant, Shanqing Cai, Cory McLean and Abhinay Ramparasad. The static version of this manuscript uses a template made by Ricardo Henriques.

TS receives funding from the Wellcome Trust through a Sir Henry Wellcome Postdoctoral Fellowship (210918/Z/18/Z). This work was also supported by the Francis Crick Institute which receives its core funding from Cancer Research UK (FC001043), the UK Medical Research Council (FC001043), and the Wellcome Trust (FC001043). This research was funded in whole, or in part, by the Wellcome Trust [FC001043]. For the purpose of Open Access, the authors have applied a CC BY public copyright licence to any Author Accepted Manuscript version arising from this submission.

AUTHOR CONTRIBUTIONS TS was responsible for conception, data ingestion, code, analysis, visualisation, UI, paper writing. MLB was responsible for visualization, analysis, code, CLI, open sourcing, testing, paper writing. DB was responsible for paper writing, analysis of results, code reviews. LJC was responsible for problem conception and framing, analysis of results, paper writing.

1. UniProt Consortium . Statistics: UniProtKB/TrEMBL 2019- 02 . https://www.uniprot. org/statistics/TrEMBL. Accessed: 2019 -03-30. 2. The UniProt Consortium . UniProt: a worldwide hub of protein knowledge . Nucleic Acids Research , 47 ( D1 ): D506 - D515 , 11 2018 . ISSN 0305- 1048 . doi: 10 .1093/nar/gky1049. 3. S. F. Altschul , W. Gish , W. Miller , E. W. Myers , and D. J. Lipman . Basic local alignment search tool . J. Mol. Biol ., 215 ( 3 ): 403 - 410 , Oct 1990 . 4. Amos Bairoch . Prosite: a dictionary of sites and patterns in proteins . Nucleic acids research , 19 ( Suppl ): 2241 , 1991 . 5. Anders Krogh , Michael Brown , I Saira Mian , Kimmen Sjölander, and David Haussler. Hidden markov models in computational biology: Applications to protein modeling . Journal of molecular biology , 235 ( 5 ): 1501 - 1531 , 1994 . 6. S R Eddy. Profile hidden Markov models . Bioinformatics , 14 ( 9 ): 755 - 763 , 10 1998 . ISSN 1367- 4803 . doi: 10 .1093/bioinformatics/14.9.755. 7. Alex Bateman , Alfredo Smart, Aurélien Luciani, Gustavo A Salazar, Jaina Mistry, Lorna J Richardson, Matloob Qureshi, Sara El-Gebali, Simon C Potter, Robert D Finn , Sean R Eddy, Erik L L Sonnhammer , Damiano Piovesan, Lisanna Paladin, Silvio C E Tosatto , and Layla Hirsh. The Pfam protein families database in 2019 . Nucleic Acids Research , 47 ( D1 ): D427 - D432 , 10 2018 . ISSN 0305- 1048 . doi: 10 .1093/nar/gky995. 8. Matthias Blum , Hsin-Yu Chang , Sara Chuguransky, Tiago Grego, Swaathi Kandasaamy, Alex Mitchell, Gift Nuka, Typhaine Paysan-Lafosse, Matloob Qureshi, Shriya Raj , et al. The interpro protein families and domains database: 20 years on . Nucleic Acids Research , 49 ( D1 ): D344 - D354 , 2021 . 9. Stephen F Altschul, Thomas L Madden, Alejandro A Schäffer, Jinghui Zhang , Zheng Zhang, Webb Miller, and David J Lipman. Gapped blast and psi-blast: a new generation of protein database search programs . Nucleic acids research , 25 ( 17 ): 3389 - 3402 , 1997 . 10. S. R. Eddy . Accelerated Profile HMM Searches . PLoS Comput. Biol ., 7 ( 10 ):e1002195, Oct 2011 . 11. Johannes Söding . Protein homology detection by hmm-hmm comparison . Bioinformatics , 21 ( 7 ): 951 - 960 , 2005 . 12. Martin Steinegger , Markus Meier, Milot Mirdita, Harald Vöhringer, Stephan J Haunsberger , and Johannes Söding . Hh-suite3 for fast remote homology detection and deep protein annotation . BMC bioinformatics , 20 ( 1 ): 1 - 15 , 2019 . 13. Mark Johnson, Irena Zaretskaya, Yan Raytselis, Yuri Merezhuk , Scott McGinnis , and Thomas L Madden. Ncbi blast: a better web interface . Nucleic acids research , 36 ( suppl . 2): W5 - W9 , 2008 . 14. Johannes Söding , Andreas Biegert, and Andrei N Lupas. The hhpred interactive server for protein homology detection and structure prediction . Nucleic acids research , 33 ( suppl . 2): W244 - W248 , 2005 . 15. Simon C Potter , Aurélien Luciani, Sean R Eddy, Youngmi Park, Rodrigo Lopez, and Robert D Finn . Hmmer web server: 2018 update . Nucleic acids research , 46 ( W1 ): W200 - W204 , 2018 . 16. Michael Bernhofer , Christian Dallago, Tim Karl, Venkata Satagopam, Michael Heinzinger, Maria Littmann, Tobias Olenyi, Jiajun Qiu, Konstantin Schuetze, Guy Yachdav , et al. Predictprotein-predicting protein structure and function for 29 years . bioRxiv, 2021 . 17. M. N. Price , K. M. Wetmore , R. J. Waters , M. Callaghan , J. Ray , H. Liu, J. V. Kuehl , R. A. Melnyk , J. S. Lamson , Y. Suh , H. K. Carlson , Z. Esquivel , H. Sadeeshkumar , R. Chakraborty , G. M. Zane , B. E. Rubin , J. D. Wall , A. Visel , J. Bristow , M. J. Blow , A. P. Arkin , and A. M. Deutschbauer . Mutant phenotypes for thousands of bacterial genes of unknown function . Nature , 557 ( 7706 ): 503 - 509 , 05 2018 . 18. C. Ramakrishnan , V. S. Dani , and T. Ramasarma . A conformational analysis of Walker motif A [GXXXXGKT (S)] in nucleotide-binding and other proteins . Protein Eng ., 15 ( 10 ): 783 - 798 , Oct 2002 . 19. S El-Gebali , L Richardson , and R Finn . Creating protein families , 2018 . 20. S El-Gebali , L Richardson , and R Finn . Repeats in pfam, 2018 . 21. Shan Carter , Zan Armstrong, Ludwig Schubert, Ian Johnson, and Chris Olah . Activation atlas . Distill , 2019 . doi: 10 .23915/distill.00015. https://distill.pub/2019/activation-atlas. 22. A. Dalkiran , A. S. Rifaioglu , M. J. Martin , R. Cetin-Atalay , V. Atalay , and T. Do ?an. ECPred: a tool for the prediction of the enzymatic functions of protein sequences based on the EC nomenclature . BMC Bioinformatics , 19 ( 1 ): 334 , Sep 2018 . 23. Maxat Kulmanov , Mohammed Asif Khan, and Robert Hoehndorf. DeepGO: predicting protein functions from sequence and interactions using a deep ontology-aware classifier . Bioinformatics , 34 ( 4 ): 660 - 668 , 10 2017 . ISSN 1367- 4803 . doi: 10 .1093/bioinformatics/btx624. 24. R. Cao , C. Freitas , L. Chan , M. Sun , H. Jiang , and Z. Chen . ProLanGO: Protein Function Prediction Using Neural Machine Translation Based on a Recurrent Neural Network. Molecules , 22 ( 10 ), Oct 2017 . 25. J. J. Almagro Armenteros , C. K. S ?nderby, S. K. S?nderby , H. Nielsen, and O. Winther. DeepLoc: prediction of protein subcellular localization using deep learning . Bioinformatics , 33 ( 21 ): 3387 - 3395 , Nov 2017 . 26. Ariel S Schwartz , Gregory J Hannum, Zach R Dwiel, Michael E Smoot, Ana R Grant, Jason M Knight , Scott A Becker , Jonathan R Eads, Matthew C LaFave, Harini Eavani, Yinyin Liu, Arjun K Bansal, and Toby H Richardson . Deep semantic protein representation for annotation, discovery, and engineering . bioRxiv, 2018 . doi: 10 .1101/365965. 27. A. Sureyya Rifaioglu , T. Do?an, M. Jesus Martin, R. Cetin-Atalay , and V. Atalay. DEEPred: Automated Protein Function Prediction with Multi-task Feed-forward Deep Neural Networks . Sci Rep , 9 ( 1 ): 7344 , May 2019 . 28. Y. Li , S. Wang , R. Umarov , B. Xie , M. Fan , L. Li , and X. Gao . DEEPre: sequence-based enzyme EC number prediction by deep learning . Bioinformatics , 34 ( 5 ): 760 - 769 , 03 2018 . 29. J. Hou , B. Adhikari , and J. Cheng. DeepSF: deep convolutional neural network for mapping protein sequences to folds . Bioinformatics, 34 ( 8 ): 1295 - 1303 , 04 2018 . 30. Maria Littmann , Michael Heinzinger , Christian Dallago, Tobias Olenyi, and Burkhard Rost . Embeddings from deep learning transfer go annotations beyond homology . Scientific reports , 11 ( 1 ): 1 - 14 , 2021 . 31. Mohammed AlQuraishi. End-to-end differentiable learning of protein structure . Cell systems , 8 ( 4 ): 292 - 301 , 2019 . 32. Andrew W Senior, Richard Evans, John Jumper, James Kirkpatrick, Laurent Sifre, Tim Green, Chongli Qin, Augustin Zidek, Alexander WR Nelson, Alex Bridgland , et al. Improved protein structure prediction using potentials from deep learning . Nature , 577 ( 7792 ): 706 - 710 , 2020 . 33. Jianyi Yang , Ivan Anishchenko, Hahnbeom Park, Zhenling Peng, Sergey Ovchinnikov, and David Baker . Improved protein structure prediction using predicted interresidue orientations . Proceedings of the National Academy of Sciences , 117 ( 3 ): 1496 - 1503 , 2020 . ISSN 0027- 8424 . doi: 10 .1073/pnas.1914677117. 34. Jianyi Yang , Ivan Anishchenko, Hahnbeom Park, Zhenling Peng, Sergey Ovchinnikov, and David Baker . Improved protein structure prediction using predicted interresidue orientations . Proceedings of the National Academy of Sciences , 117 ( 3 ): 1496 - 1503 , 2020 . 35. Yilun Du , Joshua Meier, Jerry Ma, Rob Fergus, and Alexander Rives . Energy-based models for atomic-resolution protein conformations . In International Conference on Learning Representations , 2019 . 36. Roshan Rao , Jason Liu, Robert Verkuil, Joshua Meier, John F Canny, Pieter Abbeel , Tom Sercu, and Alexander Rives . Msa transformer . bioRxiv, 2021 . 37. Adam J Riesselman , John B Ingraham, and Debora S Marks. Deep generative models of genetic variation capture the effects of mutations . Nature methods , 15 ( 10 ): 816 - 822 , 2018 . 38. Alexander Rives , Joshua Meier, Tom Sercu, Siddharth Goyal, Zeming Lin, Jason Liu, Demi Guo, Myle Ott, C Lawrence Zitnick , Jerry Ma, et al. Biological structure and function emerge from scaling unsupervised learning to 250 million protein sequences . Proceedings of the National Academy of Sciences , 118 ( 15 ), 2021 . 39. Roshan Rao , Nicholas Bhattacharya, Neil Thomas, Yan Duan , Xi Chen, John Canny, Pieter Abbeel, and Yun S Song. Evaluating protein transfer learning with tape . Advances in Neural Information Processing Systems , 32 : 9689 , 2019 . 40. Jonathan Frazer , Pascal Notin, Mafalda Dias, Aidan Gomez, Kelly Brock, Yarin Gal, and Debora Marks . Large-scale clinical interpretation of genetic variants using evolutionary data and deep learning . bioRxiv, 2020 . 41. Kevin K Yang , Zachary Wu , and Frances H Arnold . Machine-learning-guided directed evolution for protein engineering . Nature methods , 16 ( 8 ): 687 - 694 , 2019 . 42. Stanislav Mazurenko , Zbynek Prokop, and Jiri Damborsky . Machine learning in enzyme engineering . ACS Catalysis , 10 ( 2 ): 1210 - 1223 , 2019 . 43. Surojit Biswas , Grigory Khimulya, Ethan C Alley, Kevin M Esvelt , and George M Church. Low-n protein engineering with data-efficient deep learning . BioRxiv , 2020 . 44. Ali Madani , Bryan McCann , Nikhil Naik , Nitish Shirish Keskar, Namrata Anand, Raphael R Eguchi, Po-Ssu Huang , and Richard Socher. Progen: Language modeling for protein generation . arXiv preprint arXiv:2004.03497 , 2020 . 45. Ahmed Elnaggar , Michael Heinzinger , Christian Dallago, Ghalia Rihawi, Yu Wang , Llion Jones, Tom Gibbs, Tamas Feher, Christoph Angerer, Martin Steinegger , et al. Prottrans: towards cracking the language of life's code through self-supervised deep learning and high performance computing . arXiv preprint arXiv:2007.06225 , 2020 . 46. Ivan Anishchenko , Tamuka Martin Chidyausiku, Sergey Ovchinnikov, Samuel J Pellock , and David Baker. De novo protein design by deep network hallucination . bioRxiv, 2020 . 47. Drew H Bryant , Ali Bashir, Sam Sinai, Nina K Jain, Pierce J Ogden , Patrick F Riley, George M Church , Lucy J Colwell , and Eric D Kelsic . Deep diversification of an aav capsid protein by machine learning . Nature Biotechnology , pages 1 - 6 , 2021 . 48. Nadav Brandes , Dan Ofer, Yam Peleg, Nadav Rappoport, and Michal Linial . Proteinbert: A universal deep-learning model of protein sequence and function . bioRxiv, 2021 . doi: 10 .1101/ 2021 .05.24.445464. 49. Maxwell L Bileschi , David Belanger, Drew H Bryant, Theo Sanderson, Brandon Carter , D. Sculley , Mark L DePristo, and Lucy J Colwell. Using deep learning to annotate the protein universe . bioRxiv, 2019 . doi: 10 .1101/626507. 50. Kaiming He , Xiangyu Zhang, Shaoqing Ren, and Jian Sun . Deep residual learning for image recognition . CoRR, abs/1512.03385 , 2015 . 51. Fisher Yu and Vladlen Koltun. Multi-scale context aggregation by dilated convolutions . CoRR, abs/1511.07122 , 2015 . 52. David Dohan, Andreea Gane , Maxwell Bileschi, David Belanger, and Lucy Colwell . Improving protein function annotation via unsupervised pre-training: Robustness, efficiency, and insights . In Proceedings of the ACM SIGKDD International Conference on Knowledge Discovery and Data Mining , 2021 . 53. Predrag Radivojac , Wyatt T Clark, Tal Ronnen Oron, Alexandra M Schnoes, Tobias Wittkop , Artem Sokolov, Kiley Graim, Christopher Funk, Karin Verspoor, Asa Ben-Hur , et al. A largescale evaluation of computational protein function prediction . Nature methods , 10 ( 3 ): 221 - 227 , 2013 . 54. Diederik P Kingma and Jimmy Ba . Adam: A method for stochastic optimization . The International Conference on Learning Representations , 2015 . 55. Jeffrey Dean , Greg Corrado , Rajat Monga, Kai Chen, Matthieu Devin, Mark Mao, Marc'aurelio Ranzato , Andrew Senior, Paul Tucker, Ke Yang , et al. Large scale distributed deep networks . In Advances in neural information processing systems , pages 1223 - 1231 , 2012 . 56. Christopher J. Shallue , Jaehoon Lee , Joseph M. Antognini , Jascha Sohl-Dickstein, Roy Frostig , and George E. Dahl . Measuring the effects of data parallelism on neural network training . CoRR , abs/ 1811 .03600, 2018 . 57. Julian Gough , Arun Prasad Pandurangan, Ben Smithers, Jonathan Stahlhacke, and Matt E Oates. The SUPERFAMILY 2 . 0 database: a significant proteome update and a new webserver . Nucleic Acids Research , 47 ( D1 ): D490 - D494 , 11 2018 . ISSN 0305- 1048 . doi: 10 .1093/nar/gky1130. 58. T. K. Attwood , P. Bradley , D. R. Flower , A. Gaulton , N. Maudling , A. L. Mitchell , G. Moulton, A. Nordle , K. Paine , P. Taylor , A. Uddin, and C. Zygouri. PRINTS and its automatic supplement, prePRINTS . Nucleic Acids Res ., 31 ( 1 ): 400 - 402 , Jan 2003 . 59. D. H. Haft , J. D. Selengut , R. A. Richter , D. Harkins , M. K. Basu , and E. Beck. TIGRFAMs and Genome Properties in 2013 . Nucleic Acids Res ., 41 (Database issue): D387 - 395 , Jan 2013 . 60. H. Mi, S. Poudel , A. Muruganujan , J. T. Casagrande , and P. D. Thomas. PANTHER version 10: expanded protein families and functions, and analysis tools . Nucleic Acids Res ., 44 ( D1 ): D336 - 342 , Jan 2016 . 61. S. Hunter , R. Apweiler , T. K. Attwood , A. Bairoch , A. Bateman , D. Binns , P. Bork , U. Das , L. Daugherty , L. Duquenne , R. D. Finn , J. Gough , D. Haft , N. Hulo , D. Kahn , E. Kelly , A. Laugraud , I. Letunic , D. Lonsdale , R. Lopez , M. Madera , J. Maslen , C. McAnulla , J. McDowall , J. Mistry , A. Mitchell, N. Mulder , D. Natale , C. Orengo , A. F. Quinn , J. D. Selengut , C. J. Sigrist , M. Thimma , P. D. Thomas , F. Valentin , D. Wilson, C. H. Wu , and C. Yeats . InterPro: the integrative protein signature database . Nucleic Acids Res ., 37 (Database issue): D211 - 215 , Jan 2009 . 62. Baris E Suzek , Yuqi Wang , Hongzhan Huang, Peter B McGarvey , Cathy H Wu , and UniProt Consortium . Uniref clusters: a comprehensive and scalable alternative for improving sequence similarity searches . Bioinformatics , 31 ( 6 ): 926 - 932 , 2015 . 63. Naihui Zhou , Yuxiang Jiang, Timothy R Bergquist, Alexandra J Lee, Balint Z Kacsoh , Alex W Crocker, Kimberley A Lewis, George Georghiou, Huy N Nguyen, Md Nafiz Hamid , et al. The cafa challenge reports improved protein function prediction and new functional annotations for hundreds of genes through experimental screens . Genome biology , 20 ( 1 ): 1 - 23 , 2019 . 64. J. Gillis and P. Pavlidis . Characterizing the state of the art in the computational assignment of gene function: lessons from the first critical assessment of functional annotation (CAFA) . BMC Bioinformatics , 14 Suppl 3 : S15 , 2013 . 65. Alex Warwick Vesztrocy and Christophe Dessimoz . Benchmarking gene ontology function predictions using negative annotations . Bioinformatics , 36 ( Supplement _1): i210 - i218 , 2020 . 66. Martin Abadi , Paul Barham, Jianmin Chen, Zhifeng Chen, Andy Davis, Jeffrey Dean , Matthieu Devin , Sanjay Ghemawat, Geoffrey Irving, Michael Isard , et al. Tensorflow: A system for large-scale machine learning . In 12th USENIX symposium on operating systems design and implementation (OSDI 16) , pages 265 - 283 , 2016 . 67. Sean R Eddy. Accelerated profile hmm searches . PLoS computational biology , 7 ( 10 ): e1002195 , 2011 . 68. Bolei Zhou , Aditya Khosla, Agata Lapedriza, Aude Oliva, and Antonio Torralba. Learning deep features for discriminative localization . CoRR, abs/1512.04150 , 2015 . 69. Leland McInnes , John Healy, Nathaniel Saul, and Lukas Grossberger . Umap: Uniform manifold approximation and projection . The Journal of Open Source Software , 3 ( 29 ): 861 , 2018 . 70. D. Y. Chao, Y. Chen , J. Chen , S. Shi , Z. Chen , C. Wang , J. M. Danku , F. J. Zhao , and D. E. Salt . Genome-wide association mapping identifies a new arsenate reductase enzyme critical for limiting arsenic accumulation in plants . PLoS Biol , 12 ( 12 ):e1002009, Dec 2014 . 71. A. Bartels , H. P. Mock , and J. Papenbrock. Differential expression of Arabidopsis sulfurtransferases under various growth conditions . Plant Physiol Biochem , 45 ( 3-4 ): 178 - 187 , 2007 . 72. Ethan C Alley , Grigory Khimulya, Surojit Biswas, Mohammed AlQuraishi, and George M Church. Unified rational protein engineering with sequence-based deep representation learning . Nature methods , 16 ( 12 ): 1315 - 1322 , 2019 . 73. Daniel Smilkov, Nikhil Thorat, Yannick Assogba, Ann Yuan, Nick Kreeger, Ping Yu, Kangyi Zhang, Shanqing Cai, Eric Nielsen, David Soergel, Stan Bileschi , Michael Terry, Charles Nicholson, Sandeep N. Gupta , Sarah Sirajuddin , D. Sculley , Rajat Monga, Greg Corrado, Fernanda B. Viegas , and Martin Wattenberg . Tensorflow.js: Machine learning for the web and beyond, 2019 . 74. Philip Jones , David Binns, Hsin-Yu Chang , Matthew Fraser, Weizhong Li , Craig McAnulla , Hamish McWilliam , John Maslen, Alex Mitchell, Gift Nuka, Sebastien Pesseat, Antony F. Quinn, Amaia Sangrador-Vegas, Maxim Scheremetjew, Siew-Yit Yong , Rodrigo Lopez, and Sarah Hunter. InterProScan 5: genome-scale protein function classification . Bioinformatics , 30 ( 9 ): 1236 - 1240 , 01 2014 . ISSN 1367- 4803 . doi: 10 .1093/bioinformatics/btu031. 75. Michael Ashburner , Catherine A Ball, Judith A Blake, David Botstein , Heather Butler , J Michael Cherry , Allan P Davis, Kara Dolinski, Selina S Dwight, Janan T Eppig , et al. Gene ontology: tool for the unification of biology . Nature genetics , 25 ( 1 ): 25 - 29 , 2000 . 76. Gene Ontology Consortium. The gene ontology resource: 20 years and still going strong . Nucleic acids research , 47 ( D1 ): D330 - D338 , 2019 . 77. Seth Carbon , Amelia Ireland, Christopher J Mungall, ShengQiang Shu , Brad Marshall, Suzanna Lewis, AmiGO Hub, and Web Presence Working Group. Amigo: online access to ontology and annotation data . Bioinformatics , 25 ( 2 ): 288 - 289 , 2009 . 78. Yuxiang Jiang , Wyatt T Clark, Iddo Friedberg, and Predrag Radivojac. The impact of incomplete knowledge on the evaluation of protein function prediction: a structured-output learning perspective . Bioinformatics , 30 ( 17 ): i609 - i616 , 2014 . 79. L. Li , C. J. Stoeckert , and D. S. Roos . OrthoMCL: identification of ortholog groups for eukaryotic genomes . Genome Res , 13 ( 9 ): 2178 - 2189 , Sep 2003 . 80. Chuan Guo , Geoff Pleiss, Yu Sun , and Kilian Q. Weinberger . On calibration of modern neural networks . CoRR, abs/1706.04599 , 2017 . 81. Dario Amodei , Chris Olah, Jacob Steinhardt, Paul F. Christiano , John Schulman, and Dan Mané . Concrete problems in AI safety . CoRR, abs/1606.06565 , 2016 . 82. Alex Mitchell, Hsin-Yu Chang , Louise Daugherty, Matthew Fraser, Sarah Hunter, Rodrigo Lopez, Craig McAnulla , Conor McMenamin , Gift Nuka , Sebastien Pesseat , et al. The interpro protein families database: the classification resource after 15 years . Nucleic acids research , 43 ( D1 ): D213 - D221 , 2015 .