Masked Language Modeling for Proteins via Linearly Scalable LongContext Transformers
Abstract
Transformer models have achieved stateoftheart results across a diverse range of domains. However, concern over the cost of training the attention mechanism to learn complex dependencies between distant inputs continues to grow. In response, solutions that exploit the structure and sparsity of the learned attention matrix have blossomed. However, realworld applications that involve long sequences, such as biological sequence analysis, may fall short of meeting these assumptions, precluding exploration of these models. To address this challenge, we present a new Transformer architecture, , based on Fast Attention Via Orthogonal Random features (FAVOR). Our mechanism scales linearly rather than quadratically in the number of tokens in the sequence, is characterized by subquadratic space complexity and does not incorporate any sparsity pattern priors. Furthermore, it provides strong theoretical guarantees: unbiased estimation of the attention matrix and uniform convergence. It is also backwardscompatible with pretrained regular Transformers. We demonstrate its effectiveness on the challenging task of protein sequence modeling and provide detailed theoretical analysis.
1 Introduction and related work
Transformers Vaswani et al. (2017); Dehghani et al. (2019) are powerful neural network architectures that have become SOTA in several areas of machine learning including Natural Language Processing (NLP) (e.g. speech recognition Luo et al. (2020)), Neural Machine Translation (NMT) Chen et al. (2018), document generation/summarization, time series prediction, generative modeling (e.g. image generation Parmar et al. (2018)), music generation Huang et al. (2019), and analysis of biological sequences Rives et al. (2019); Madani et al. (2020); Li (2019). Transformers rely on a trainable attention mechanism that specifies complex dependencies between the elements of each input sequence (e.g. amino acids within a protein). Unfortunately, a standard Transformer scales quadratically with the number of tokens in the input sequence, which is prohibitively expensive for large . Several solutions have been proposed to address this issue Beltagy et al. (2020); Gulati et al. (2020); Chan et al. (2020); Child et al. (2019). Most approaches restrict the attention mechanism to attend to local neighborhoods Parmar et al. (2018) or incorporate structural priors on attention such as sparsity Child et al. (2019), poolingbased compression Rae et al. (2020) clustering/binning/convolution techniques (e.g. Roy et al. (2020) which applies means clustering to learn dynamic sparse attention regions, or Kitaev et al. (2020), where locality sensitive hashing is used to group together tokens of similar embeddings), sliding windows Beltagy et al. (2020), or truncated targeting Chelba et al. (2020). Thus these approaches do not aim to approximate regular attention, but rather propose simpler and more tractable attention mechanisms, often by incorporating additional constraints (e.g. identical query and key sets as in Kitaev et al. (2020)), or by trading regular attention with sparse attention using more layers Child et al. (2019). Furthermore, many of these works require special customized GPU operations (e.g. either writing C++ CUDA kernels Child et al. (2019) or using TVMs Beltagy et al. (2020)). Other techniques which aim to improve the time complexity of Transformers include reversible residual layers allowing for onetime activation storage in training Kitaev et al. (2020) and shared attention weights Xiao et al. (2019). These constraints may impede application to problems that involve long sequences, where approximations of the attention mechanism are not sufficient. Approximations based on truncated backpropagation Dai* et al. (2019) are also unable to capture longdistance correlations since the gradients are only propagated inside a localized window.
Recent work has demonstrated that Transformers fit to the amino acid sequences of single proteins learn to accurately predict information about protein structure and function, and can generate new sequences with specific properties Rives et al. (2019); Elnaggar et al. (2019); Madani et al. (2020). Approaches that encode 3D protein structural data via Transformerbased models demonstrate improved performance, despite the restriction of attention to the local structural neighborhoods of each node Du et al. (2020); Ingraham et al. (2019). These models provide initial promise for protein design applications, but their applicability beyond the design of single proteins is limited because they truncate sequences to 512 or 1024 amino acids. The ability to scale to longer sequences without imposing sparsity constraints would enable the use of Transformers to jointly model multiple concatenated protein sequences and the interactions between them. This follows recent works employing simpler statistical models that predict protein quaternary structure, proteinprotein interactions and protein interaction networks from evolutionary sequence data Weigt et al. (2009); Hopf et al. (2012); Ovchinnikov et al. (2014); Bitbol et al. (2016); Cong et al. (2019).
In response, we present a new Transformer architecture, , based on Fast Attention Via Orthogonal Random features (FAVOR). Our proposed mechanism has several properties required by modern protein modeling: it scales linearly rather than quadratically in the number of tokens in the sequence (important for analysis involving compounds of protein molecules), is characterized by subquadratic space complexity, and does not incorporate any sparsity patterns priors. Furthermore, it provides strong theoretical guarantees: unbiased estimation of the regular attention matrix and uniform convergence. FAVOR is designed for long input sequences where the number of tokens satisfies , for embedding dimensionality . In contrast to previous approaches, instead of simplifying regular attention via various structural priors (which can lead to different, potentially incompatible architectures), we show that it can be effectively approximated as it is, without any "liftings". This leads to our method being flexible: combined with small amounts of finetuning, the Performer is backwardscompatible with pretrained regular Transformers and can be also used beyond the Transformer scope as a more scalable replacement of regular attention, which itself has a wide variety of uses in computer vision Fu et al. (2019), reinforcement learning Zambaldi et al. (2019), and even combinatorial optimization Vinyals et al. (2015). We demonstrate its effectiveness on the challenging task of protein modeling.
We show that regular attention can be considered a special case of a much larger class of kerneldriven attention mechanisms, Generalized Attention (GA), and that all our results for regular attention can be directly translated also to this extended class. This observation enables us to explore a much larger class of attention models (Sec. 2.2). Interestingly, this is often enabled by the FAVOR mechanism, even if linear scaling is not required (Sec. 4). We highlight the following contributions:

We present a general class of kernelbased attention mechanisms, Generalized Attention (GA), which can be handled by FAVOR. Standard attention is a special case. (Sec. 2.2).

We empirically evaluate FAVOR via for protein modeling, demonstrating in practice all the aforementioned advantages (Sec. 4).

We show that our mechanism, implemented in Jax Frostig et al. (2018), is APIcompatible with the regular Transformer, whose standard dotproduct attention can be replaced by FAVOR with all other components of the architecture intact.
All proofs are given in full in the Appendix.
2 Generalized Attention via FAVOR mechanism
Below we describe in detail our FAVOR mechanism which is the backbone of our architecture. We also present a general class of kernelbased attentions, called Generalized Attention (GA) (which includes regular attention as a special case), where FAVOR can be applied.
2.1 Preliminaries  standard attention mechanism
Let be the size of an input sequence of tokens. Then regular dotproduct attention Vaswani et al. (2017) is a mapping which accepts matrices as input where is the hidden dimension (dimension of the latent representation). Matrices are intermediate representations of the input and their rows can be interpreted as queries, keys and values of the continuous dictionary data structure respectively. Bidirectional (or nondirectional Devlin et al. (2018)) dotproduct attention has the following form:
(1) 
where is applied elementwise, is the allones vector of length , and is a diagonal matrix with the input vector as the diagonal. The runtime complexity of computing (1) is because the attention matrix has to be computed and stored explicitly. Hence, in principle, dotproduct attention of type (1) is incompatible with endtoend processing of long sequences.
Another important type of attention is unidirectional dotproduct attention which has the form:
(2) 
where returns the lowertriangular part of the argument matrix including diagonal. As discussed in Vaswani et al. (2017), unidirectional attention is used for autoregressive generative modelling with Transformers when the output sequence is modelled as:
Therefore, the probability distribution over can only depend on embeddings of tokens . Unidirectional attention is used as selfattention in generative Transformers as well as the decoder part of Seq2Seq Transformers Vaswani et al. (2017), while bidirectional attention is used in encoder selfattention and encoderdecoder attention in Seq2Seq architectures.
A line of work relies on sparse approximation of the matrix – either through restricting the sparsity pattern of Child et al. (2019) or learning it using LocalitySensitive Hashing (LSH) techniques Kitaev et al. (2020). The latter results in runtime complexity. We will show that, without any structural assumptions, the matrix can be approximated up to any precision in time .
2.2 Generalized Attention (GA)
The idea of the attention mechanism is simple. New representations of tokens are obtained from previous ones by taking convex combinations of different value vectors with coefficients of the convex combinations interpreted as renormalized (i.e. all coefficients sum up to one) similarity measures between different tokens. High similarities imply strong attendance to the corresponding tokens. These similarity measures are simple adhoc “softmax style" functions of a dotproduct between query of token and key of token , namely:
(3) 
where: . Note that is not a commutative operation here, and the renormalizer is a technical modification to stabilize the range of and avoid very small/large values.
However, what if we use kernels instead of arbitrary similarity measures? Specifically, and are entangled through a valid kernel function, by defining the attention matrix as:
(4) 
where is an arbitrary kernel function and . We call this attention mechanism defined above Generalized Attention (GA) parameterized by .
Next we show that not only can FAVOR approximate regular attention governed by Eq. 3, but it can be applied to GAs as long as the corresponding kernels can be effectively estimated via a random feature map mechanism Rahimi and Recht (2007), which is the case for most kernels used in practice. We will in fact show that regular attention is a special case of GA for a specific choice of , and Gaussian kernel .
2.3 Towards FAVOR: approximating attention with random features (RFs)
Instead of computing and storing the attention matrix explicitly, we derive its unbiased stochastic approximation, which benefits from lowrank structure. We take our inspiration from a randomized scheme to train kernel Support Vector Machines with large training data Rahimi and Recht (2007).
Let and denote the th rows of matrices and respectively. For regular attention, the th element of can be expressed as:
In other words, for , the attention matrix can be decomposed as:
(5)  
(6) 
for . Both and can be computed in time. Note that the th element of matrix is the value of the Gaussian kernel with :
(7) 
For GA, our analysis is similar. This time have nonzero entries of the form and (for regular attention we have: ) respectively and furthermore the Gaussian kernel is replaced by a general kernel , namely: , as in Equation 4.
In the reminder of this section we will derive an unbiased stochastic approximation of matrix based on lowrank decomposition of with the use of random feature maps Rahimi and Recht (2007).
For a given kernel , the random feature [RF] map corresponding to is a probabilistic embedding satisfying
(8) 
where the expectation is with respect to the randomness of , and denotes the number of random features (if only approximates then we refer to the mechanism as an approximate random feature map). Efficienttocompute random feature maps exist for virtually all classes of kernels used in machine learning, e.g. shiftinvariant kernels Rahimi and Recht (2007), the pointwise nonlinear Gaussian kernel related to neural networks Gulrajani et al. (2017), and more, though the techniques used to derive these random mappings vary from class to class Choromanski et al. (2017). Even more interestingly, for most of these kernels, corresponding random feature maps have a similar structure, namely:
(9) 
for some , , , distributions: , and constant . Here has rows and .
In particular, for the Gaussian kernel, we have and:
(10) 
where and . This particular form of is a consequence of the celebrated Bochner’s Theorem Rahimi and Recht (2007). We now define and as:
(11) 
Note that we have: and where and stand for the ith row of and respectively. Then according to Equation 8, we have: . Thus with , given as: , , we obtain:
(12) 
We conclude that the attention matrix can be approximated without bias as: . We will leverage this unbiased approximate lowrank (if ) decomposition of in our algorithm, even though we will not explicitly compute .
Note that one can also define a valid kernel as: for as in Eq. 9 and an arbitrary . Such kernels cover in particular the family of Pointwise Nonlinear Gaussian Kernels Choromanski et al. (2017) (intrinsically related to nonlinear neural networks) such as arccosine kernels (e.g. angular kernels). Most of these kernels do not have closedforms so computing exact GAs for them would not be possible, but of course computation is feasible with the presented mechanism.
2.4 Towards FAVOR: refinements via orthogonal random features
For isotropic (true for most practical applications, including regular attention), instead of sampling independently, we can use orthogonal random features (ORF) Yu et al. (2016); Choromanski et al. (2017, 2018b): these maintain (exactly or approximately) the marginal distributions of samples while enforcing that different samples are orthogonal. If we need , ORFs still can be used locally within each block of Yu et al. (2016).
ORFs were introduced to reduce the variance of Monte Carlo estimators Yu et al. (2016); Choromanski et al. (2017, 2018b, 2019a); Rowland et al. (2019); Choromanski et al. (2018a, 2019b) and we show in Secs. 3 and 4 that they do indeed lead to more accurate approximations and substantially better downstream results. Below we breifly review the most efficient ORF mechanisms (based on their strengths and costs) that we will use in Sec. 2.6 in the analysis of FAVOR.
(1) Regular ORFs [RORFs]: Applies Gaussian orthogonal matrices Yu et al. (2016). Encodes matrix in space. Provides algorithm for computing in time for any . Gives unbiased estimation. Requires onetime preprocessing (GramSchmidt orthogonalization).
2.5 FAVOR: Fast Attention via Orthogonal Random features
We are ready to present the full FAVOR algorithm. In the bidirectional case, our approximate attention computed by FAVOR is given as:
(13) 
where . The placement of brackets determines the order in which computations are conducted. Note that we never explicitly compute and consequently, avoid time complexity and storing the approximate attention matrix (see: Sec. 2.6 for rigorous analysis).
2.5.1 Prefixsums for unidirectional FAVOR
For the unidirectional case, our analysis is similar but this time our goal is to compute without constructing and storing the sized matrix explicitly, where : . In order to do so, observe that
(14) 
where are 3dtensors. Each slice is therefore a result of a prefixsum (or cumulativesum) operation applied to : . An efficient algorithm to compute the prefixsum of elements takes total steps and time when computed in parallel Ladner and Fischer (1980); Cormen et al. (2009). See Algorithm 1 for the whole approach.
2.6 Time and space complexity analysis
We see that a variant of bidirectional FAVOR using regular RFs (based on iid samples) or RORFs has space complexity as opposed to space complexity of the baseline. Unidirectional FAVOR using fast prefixsum precomputation in parallel Ladner and Fischer (1980); Cormen et al. (2009) has space complexity to store which can be reduced to by running a simple (though nonparallel in ) aggregation of without storing the whole tensor in memory. From Sec. 2.4, we know that if instead we use GORFs, then space complexity is reduced to and if the HORFs mechanism is used, then space is further reduced to . Thus for all our variants provide substantial space complexity improvements since they do not need to store the attention matrix explicitly.
The time complexity of Algorithm 1 is (note that constructing and can be done in time via Eq. 11 if samples from and can be obtained in time and respectively (which is the case for all practical applications). Note that the time complexity of our method is much lower than of the baseline for .
As explained in Sec. 2.4, the RORF mechanism incurs an extra onetime cost (negligible compared to the term for ). HORFs or GORFs do not have this cost, and when FAVOR uses them, computing and can be conducted in time as opposed to (see: Sec. 2.4). Thus even though H/GORFs do not change the asymptotic time complexity, they improve the constant factor from the leading term. This plays an important role for training very large models.
The number of random features allows a tradeoff between computational complexity and the level of approximation: bigger results in higher computation costs, but also in a lower variance of the estimate of . In the next section we will show that in practice we can take .
Observe that the algorithm obtained is highlyparallelizable, and benefits from fast matrix multiplication and broadcasted operations on GPUs or TPUs.
3 Theoretical convergence analysis
In contrast to other methods approximating the attention matrix , our algorithm provides provable strong uniform convergence theoretical guarantees for compact domains. We show that , the optimal number of random features, does not depend on but only on . In fact, we prove that if we take , then with time, we can approximate up to any precision, regardless of the number of tokens . In order to provide those guarantees for FAVOR, we leverage recent research on the theory of negative dependence for ORFs Lin et al. (2020). The following is true:
Theorem 1 (Uniform convergence of FAVOR).
Take the generalized attention mechanism defined by (see: Sec. 2.2) and a radial basis function (RBF) kernel Choromanski et al. (2018b) with corresponding spectral distribution (e.g. Gaussian kernel for which ). Assume that the rows of matrices and are taken from a ball of radius , centered at (i.e. norms of queries and keys are upperbounded by ). Define and take , . Then for any , and the number of random features
The result holds in particular for regular attention using Gaussian kernels (see: Sec. 2.2) for which
4 Experiments
We implement our setup on top of preexisting Transformer training code in Jax Frostig et al. (2018), and complement our theory with empirical evidence to demonstrate FAVOR’s practicality in the protein setting. Unless explicitly stated, a Performer replaces only the attention component with FAVOR, while all other components are exactly the same as for the regular Transformer. Furthermore, since we use the crossentropy loss in our generative training experiments, we use standard the accuracy metric as defined from supervised learning.
4.1 Computation costs
We compared speedwise the backward pass, as it is one of the main computational bottlenecks during training, for a Transformer and a Performer in two settings: when the architecture is mostly composed of attention while other dimensions are small , as well as the regular default size , where denotes the width of the MLP layers of the Transformer. We observed (Fig. 1) that in terms of , the Performer reaches nearly linear time complexity as opposed to the Transformer’s quadratic time complexity. Furthermore, the Performer’s memory consumption is subquadratic (as it does not store the explicit sized attention matrix), which allows both higher batch sizes and longer sequence lengths. The combination of both memory and backward pass efficiencies for large has profound implications for training speed, as it allows respectively, large batch training and lower wall clock time per gradient step, contributing to total train time reduction. We present additional results, including the forward pass, in the Appendix A by varying layers and architecture sizes as well.
4.2 Approximation error and compatibility with regular Transformer
We further examined the approximation error of the attention matrix implicitly defined in FAVOR in Fig. 2 (and in Fig. 8 in Appendix B), which thus directly affects the accuracy of FAVOR’s output. We demonstrate that orthogonal features generally produce lower error than unstructured features.
Notice that the accuracy can be further boosted by applying a resampling strategy that reconstructs samples periodically. We set this period as a hyperparameter of our overall algorithm.
The approximation error can propagate when applying the other components (MLPs, multiple heads, multiple layers, etc.) of a Transformer, which we demonstrate in Fig. 7 (Appendix). This implies we cannot immediately directly transfer the weights from a pretrained Transformer onto the Performer. However, this can be resolved by finetuning the Performer on the trained task. We demonstrate this technique for a pretrained BERT model Devlin et al. (2018) on the LM1B dataset Chelba et al. (2014) in Fig. 3.
4.3 Generalized Attention
We investigated Generalized Attention mechanisms (Sec. 2.2) on protein datasets Consortium (2019) of up to length 512 for various kernel functions. Using hyperparameter sweeps across multiple variables in FAVOR, we compared several kernels and also renormalization on/off (Fig. 4, corresponds to applying operator in attention, as for the standard mechanism; though we noticed that disabling it does not necessarily hurt accuracy) to produce the best training configuration for the Performer. We found the sigmoid kernel with renormalization ON was the optimal configuration for the Performer.
4.4 Training on concatenated protein sequences
Finally, we demonstrate that the Performer can model multiple concatenated protein sequences as required to model and predict interactions among groups of proteins from sequence data. For this proof of principle study, we use protein sequences from the Jan. 2019 release of Trembl Consortium (2019), and concatenated protein sequences to length , long enough to model protein interaction networks without the large sequence alignments required by existing methods Cong et al. (2019). We train models on a Cloud TPU v3, containing 16GB RAM per chip. At this length, a baseline Transformer overloads memory even at a batch size of per chip by a wide margin. Thus as a baseline we were forced to use a significantly smaller variant, reducing to . Meanwhile, the Performer trains efficiently at a batch size of 16 per chip using the standard architecture. We see in Fig. 5 that the Transformer is quickly bounded at , while the Performer is able to train continuously, increasing its performance as training progresses.
5 Conclusion
We presented , a new type of Transformer, relying on our Fast Attention Via Orthogonal Random features (FAVOR) mechanism to significantly improve space and time complexity of regular Transformers. Our mechanism is to our knowledge the first unbiased estimation of the original algorithm with linear space and time complexity with respect to . Further, FAVOR could be applied to other tasks of approximate attention, including hierarchical attention networks (HANS) Yang et al. (2016), graph attention networks Velickovic et al. (2018), image processing Fu et al. (2019), and reinforcement learning/robotics Tang et al. (2020).
6 Broader impact
We believe that the presented algorithm can be impactful in various ways:
Biology and Medicine: Our method has the potential to directly impact research on biological sequence analysis by enabling the Transformer to be applied to much longer sequences without constraints on the structure of the attention matrix. The initial application that we consider is the prediction of interactions between proteins on the proteome scale. Recently published approaches require large evolutionary sequence alignments, a bottleneck for applications to mammalian genomes Cong et al. (2019). The potentially broad translational impact of applying these approaches to biological sequences was one of the main motivations of this work. We believe that modern bioinformatics can immensely benefit from new machine learning techniques with Transformers being among the most promising. Scaling up these methods to train faster more accurate language models opens the door to the ability to design sets of molecules with prespecified interaction properties. These approaches could be used to augment existing physicsbased design strategies that are of critical importance for example in the development of new nanoparticle vaccines Marcandalli et al. (2019).
Environment: As we have shown, Performers with FAVOR are characterized by much lower compute costs and substantially lower space complexity which can be directly translated to emission reduction Strubell et al. (2019) and lower energy consumption You et al. (2020), as regular Transformers require very large computational resources.
Research on Transformers: We believe that our results can shape research on efficient Transformers architectures, guiding the field towards methods with strong mathematical foundations. Our research may also hopefully extend Transformers also beyond their standard scope (e.g. by considering the Generalized Attention mechanism and connections with kernels). Exploring scalable Transformer architectures that can handle of the order of magnitude few thousands and more, preserving accuracy of the baseline at the same time, is a gateway to new breakthroughs in bioinformatics, e.g. language modeling for proteins, as we explained in the paper. Our presented method can be potentially a first step.
Backward Compatibility: Our Performer can be used on the top of a regular pretrained Transformer as opposed to other Transformer variants. Even if uptraining is not required, FAVOR can be still used for fast inference with no loss of accuracy. We think about this backward compatibility as a very important additional feature of the presented techniques that might be particularly attractive for practitioners.
Attention Beyond Transformers: Finally, FAVOR can be applied to approximate exact attention also outside the scope of Transformers. This opens a large volume of new potential applications including: hierarchical attention networks (HANS) Yang et al. (2016), graph attention networks Velickovic et al. (2018), image processing Fu et al. (2019), and reinforcement learning/robotics Tang et al. (2020).
References
 [1] (2020) Longformer: the longdocument transformer. CoRR abs/2004.05150. External Links: Link, 2004.05150 Cited by: §1.
 [2] (2016) Inferring interaction partners from protein sequences. Proceedings of the National Academy of Sciences 113 (43), pp. 12180–12185. Cited by: §1.
 [3] (2020) Imputer: sequence modelling via imputation and dynamic programming. CoRR abs/2002.08926. External Links: Link, 2002.08926 Cited by: §1.
 [4] (2020) Faster transformer decoding: ngram masked selfattention. CoRR abs/2001.04589. External Links: Link, 2001.04589 Cited by: §1.
 [5] (2014) One billion word benchmark for measuring progress in statistical language modeling. In INTERSPEECH 2014, 15th Annual Conference of the International Speech Communication Association, Singapore, September 1418, 2014, pp. 2635–2639. Cited by: §4.2.
 [6] (2018) The best of both worlds: combining recent advances in neural machine translation. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics, ACL 2018, Melbourne, Australia, July 1520, 2018, Volume 1: Long Papers, I. Gurevych and Y. Miyao (Eds.), pp. 76–86. External Links: Link, Document Cited by: §1.
 [7] (2019) Generating long sequences with sparse transformers. CoRR abs/1904.10509. External Links: Link, 1904.10509 Cited by: §1, §2.1.
 [8] (2018) Initialization matters: orthogonal predictive state recurrent neural networks. In 6th International Conference on Learning Representations, ICLR 2018, Vancouver, BC, Canada, April 30  May 3, 2018, Conference Track Proceedings, External Links: Link Cited by: §2.4.
 [9] (2017) The unreasonable effectiveness of structured random orthogonal embeddings. In Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, 49 December 2017, Long Beach, CA, USA, I. Guyon, U. von Luxburg, S. Bengio, H. M. Wallach, R. Fergus, S. V. N. Vishwanathan, and R. Garnett (Eds.), pp. 219–228. External Links: Link Cited by: §2.3, §2.3, §2.4, §2.4, §2.4.
 [10] (2019) KAMANNs: lowdimensional rotation based neural networks. In The 22nd International Conference on Artificial Intelligence and Statistics, AISTATS 2019, 1618 April 2019, Naha, Okinawa, Japan, K. Chaudhuri and M. Sugiyama (Eds.), Proceedings of Machine Learning Research, Vol. 89, pp. 236–245. External Links: Link Cited by: §2.4.
 [11] (2019) Unifying orthogonal Monte Carlo methods. In Proceedings of the 36th International Conference on Machine Learning, ICML 2019, 915 June 2019, Long Beach, California, USA, K. Chaudhuri and R. Salakhutdinov (Eds.), Proceedings of Machine Learning Research, Vol. 97, pp. 1203–1212. External Links: Link Cited by: §2.4, §2.4.
 [12] (2018) The geometry of random features. In International Conference on Artificial Intelligence and Statistics, AISTATS 2018, 911 April 2018, Playa Blanca, Lanzarote, Canary Islands, Spain, A. J. Storkey and F. PérezCruz (Eds.), Proceedings of Machine Learning Research, Vol. 84, pp. 1–9. External Links: Link Cited by: §2.4, §2.4, Theorem 1.
 [13] (2019) Protein interaction networks revealed by proteome coevolution. Science 365 (6449), pp. 185–189. Cited by: §1, §4.4, §6.
 [14] (2019) UniProt: a worldwide hub of protein knowledge. Nucleic acids research 47 (D1), pp. D506–D515. Cited by: §4.3, §4.4.
 [15] (2009) Introduction to algorithms, 3rd edition. MIT Press. External Links: Link, ISBN 9780262033848 Cited by: §2.5.1, §2.6.
 [16] (2019) TransformerXL: language modeling with longerterm dependency. External Links: Link Cited by: §1.
 [17] (2019) Universal transformers. In 7th International Conference on Learning Representations, ICLR 2019, New Orleans, LA, USA, May 69, 2019, External Links: Link Cited by: §1.
 [18] (2018) BERT: pretraining of deep bidirectional transformers for language understanding. CoRR abs/1810.04805. External Links: Link, 1810.04805 Cited by: §2.1, §4.2.
 [19] (2020) Energybased models for atomicresolution protein conformations. arXiv preprint arXiv:2004.13167. Cited by: §1.
 [20] (2019) Endtoend multitask learning, from protein language to protein features without alignments. bioRxiv, pp. 864405. Cited by: §1.
 [21] (2018) Compiling machine learning programs via highlevel tracing. External Links: Link Cited by: 5th item, §4.
 [22] (2019) Dual attention network for scene segmentation. In IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2019, Long Beach, CA, USA, June 1620, 2019, pp. 3146–3154. Cited by: §1, §5, §6.
 [23] (2020) Conformer: convolutionaugmented transformer for speech recognition. External Links: 2005.08100 Cited by: §1.
 [24] (2017) Improved training of Wasserstein GANs. In Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, 49 December 2017, Long Beach, CA, USA, I. Guyon, U. von Luxburg, S. Bengio, H. M. Wallach, R. Fergus, S. V. N. Vishwanathan, and R. Garnett (Eds.), pp. 5767–5777. External Links: Link Cited by: §2.3.
 [25] (2012) Threedimensional structures of membrane proteins from genomic sequencing. Cell 149 (7), pp. 1607–1621. Cited by: §1.
 [26] (2019) Music transformer: generating music with longterm structure. In 7th International Conference on Learning Representations, ICLR 2019, New Orleans, LA, USA, May 69, 2019, External Links: Link Cited by: §1.
 [27] (2019) Generative models for graphbased protein design. In Advances in Neural Information Processing Systems, pp. 15794–15805. Cited by: §1.
 [28] (2020) Reformer: the efficient transformer. In 8th International Conference on Learning Representations, ICLR 2020, Addis Ababa, Ethiopia, April 2630, 2020, External Links: Link Cited by: §1, §2.1, §A.
 [29] (198010) Parallel prefix computation. J. ACM 27 (4), pp. 831–838. External Links: ISSN 00045411, Link, Document Cited by: §2.5.1, §2.6.
 [30] (2019) Universal transforming geometric network. CoRR abs/1908.00723. External Links: Link, 1908.00723 Cited by: §1.
 [31] (2020) Demystifying orthogonal Monte Carlo and beyond. CoRR abs/2005.13590. Cited by: §3, §C.
 [32] (2020) Simplified selfattention for transformerbased endtoend speech recognition. CoRR abs/2005.10463. External Links: Link, 2005.10463 Cited by: §1.
 [33] (2020) ProGen: language modeling for protein generation. CoRR abs/2004.03497. External Links: Link, 2004.03497 Cited by: §1, §1.
 [34] (2019) Induction of potent neutralizing antibody responses by a designed protein nanoparticle vaccine for respiratory syncytial virus. Cell 176 (6), pp. 1420–1431. Cited by: §6.
 [35] (2014) Robust and accurate prediction of residue–residue interactions across protein interfaces using evolutionary information. Elife 3, pp. e02030. Cited by: §1.
 [36] (2018) Image transformer. In Proceedings of the 35th International Conference on Machine Learning, ICML 2018, Stockholmsmässan, Stockholm, Sweden, July 1015, 2018, J. G. Dy and A. Krause (Eds.), Proceedings of Machine Learning Research, Vol. 80, pp. 4052–4061. External Links: Link Cited by: §1.
 [37] (2020) Compressive transformers for longrange sequence modelling. In International Conference on Learning Representations, External Links: Link Cited by: §1.
 [38] (2007) Random features for largescale kernel machines. In Advances in Neural Information Processing Systems 20, Proceedings of the TwentyFirst Annual Conference on Neural Information Processing Systems, Vancouver, British Columbia, Canada, December 36, 2007, J. C. Platt, D. Koller, Y. Singer, and S. T. Roweis (Eds.), pp. 1177–1184. External Links: Link Cited by: §2.2, §2.3, §2.3, §2.3, §2.3, §C.
 [39] (201904) Biological structure and function emerge from scaling unsupervised learning to 250 million protein sequences. bioArxiv. External Links: Document Cited by: §1, §1.
 [40] (2019) Orthogonal estimation of Wasserstein distances. In The 22nd International Conference on Artificial Intelligence and Statistics, AISTATS 2019, 1618 April 2019, Naha, Okinawa, Japan, K. Chaudhuri and M. Sugiyama (Eds.), Proceedings of Machine Learning Research, Vol. 89, pp. 186–195. External Links: Link Cited by: §2.4.
 [41] (2020) Efficient contentbased sparse attention with routing transformers. CoRR abs/2003.05997. External Links: Link, 2003.05997 Cited by: §1.
 [42] (2019) Energy and policy considerations for deep learning in NLP. CoRR abs/1906.02243. External Links: Link, 1906.02243 Cited by: §6.
 [43] (2020) Neuroevolution of selfinterpretable agents. CoRR abs/2003.08165. External Links: Link, 2003.08165 Cited by: §5, §6.
 [44] (2017) Attention is all you need. In Advances in Neural Information Processing Systems 30, I. Guyon, U. V. Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathan, and R. Garnett (Eds.), pp. 5998–6008. External Links: Link Cited by: §1, §2.1, §2.1.
 [45] (2018) Graph attention networks. In 6th International Conference on Learning Representations, ICLR 2018, Vancouver, BC, Canada, April 30  May 3, 2018, Conference Track Proceedings, External Links: Link Cited by: §5, §6.
 [46] (2015) Pointer networks. In Advances in Neural Information Processing Systems 28: Annual Conference on Neural Information Processing Systems 2015, December 712, 2015, Montreal, Quebec, Canada, pp. 2692–2700. Cited by: §1.
 [47] (2009) Identification of direct residue contacts in protein–protein interaction by message passing. Proceedings of the National Academy of Sciences 106 (1), pp. 67–72. Cited by: §1.
 [48] (2019) Sharing attention weights for fast transformer. In Proceedings of the TwentyEighth International Joint Conference on Artificial Intelligence, IJCAI 2019, Macao, China, August 1016, 2019, S. Kraus (Ed.), pp. 5292–5298. External Links: Link, Document Cited by: §1.
 [49] (2016) Hierarchical attention networks for document classification. In NAACL HLT 2016, The 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, San Diego California, USA, June 1217, 2016, K. Knight, A. Nenkova, and O. Rambow (Eds.), pp. 1480–1489. External Links: Link, Document Cited by: §5, §6.
 [50] (2020) Drawing earlybird tickets: toward more efficient training of deep networks. In International Conference on Learning Representations, External Links: Link Cited by: §6.
 [51] (2016) Orthogonal random features. In Advances in Neural Information Processing Systems 29: Annual Conference on Neural Information Processing Systems 2016, December 510, 2016, Barcelona, Spain, D. D. Lee, M. Sugiyama, U. von Luxburg, I. Guyon, and R. Garnett (Eds.), pp. 1975–1983. Cited by: §2.4, §2.4, §2.4.
 [52] (2019) Deep reinforcement learning with relational inductive biases. In 7th International Conference on Learning Representations, ICLR 2019, New Orleans, LA, USA, May 69, 2019, Cited by: §1.
APPENDIX: Masked Language Modeling for Proteins via Linearly Scalable LongContext Transformers
A Extended computation costs
In this subsection, we empirically measure computational costs in terms wall clock time for both the forward and backward passes when we replace the attention mechanism on a regular Transformerbased architecture. Since some of the computational bottleneck in the Transformer may originate from the extra feedforward layers [28], we thus focus on the attention part of our mechanism (which is primarily dependent on ) by varying both the number of layers and sequence length, while fixing the other components to be relatively minor  i.e. with a batch size of 1.
B Extended approximation results
Although mentioned previously (Sec. 4.2) that the Performer with additional finetuning is backwards compatible with the Transformer, we demonstrate below error propagation due to nonattention components of the Transformer is one of the primary reasons that pretrained Transformer weights cannot be immediately used for inference on the corresponding Performer.
We further extend the hyperparameter sweep setting from Figure 4 in the main body of the paper, and see that across varying hyperparameters, training with orthogonal features is generally is the most accurate.
C Theoretical results
We provide here the proof of Theorem 1 from the main body.
Proof.
We consider first the case of the default FAVOR setting with RORF mechanism turned on. We rely on Theorem 3 from [31]. Note that we can apply it in our case, since for RBF kernels the corresponding function is (thus in particular it is bounded). Also, it is not hard to observe (see for instance analysis in Claim 1 from [38]) that . Using Theorem 3 from [31], we conclude that:
(15) 
with any constant probability as long as
(16) 
Taking completes the proof. ∎