Abstract
AbstractProtein language models (PLMs) have shown remarkable capabilities in various protein function prediction tasks. However, while protein function is intricately tied to structure, most existing PLMs do not incorporate protein structure information. To address this issue, we introduce ProSST, a Transformer-based protein language model that seamlessly integrates both protein sequences and structures. ProSST incorporates a structure quantization module and a Transformer architecture with disentangled attention. The structure quantization module translates a 3D protein structure into a sequence of discrete tokens by first serializing the protein structure into residue-level local structures and then embeds them into dense vector space. These vectors are then quantized into discrete structure tokens by a pre-trained clustering model. These tokens serve as an effective protein structure representation. Furthermore, ProSST explicitly learns the relationship between protein residue token sequences and structure token sequences through the sequence-structure disentangled attention. We pre-train ProSST on millions of protein structures using a masked language model objective, enabling it to learn comprehensive contextual representations of proteins. To evaluate the proposed ProSST, we conduct extensive experiments on the zero-shot mutation effect prediction and several supervised downstream tasks, where ProSST achieves the state-of-the-art performance among all baselines. Our code and pretrained models are publicly available2.
Publisher
Cold Spring Harbor Laboratory
Reference48 articles.
1. William P Jencks . Catalysis in chemistry and enzymology. Courier Corporation, 1987.
2. UniProt: the Universal Protein Knowledgebase in 2023;Nucleic Acids Research,2022
3. Language models are unsupervised multitask learners;OpenAI blog,2019
4. Bert: Pre-training of deep bidirectional transformers for language understanding;arXiv preprint,2018
5. Biological structure and function emerge from scaling unsupervised learning to 250 million protein sequences
Cited by
1 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献