FoldToken: Learning Protein Language via Vector Quantization and Beyond
DOI:
https://doi.org/10.1609/aaai.v39i1.31998Abstract
Is there a foreign language describing protein sequences and structures simultaneously? Protein structures, represented by continuous 3D points, have long posed a challenge due to the contrasting modeling paradigms of discrete sequences. We introduce FoldTokenizer to represent protein sequence-structure as discrete symbols. This approach involves projecting residue types and structures into a discrete space, guided by a reconstruction loss for information preservation. We name the learned discrete symbols as FoldToken, and the sequence of FoldTokens serves as a new protein language, transforming the protein sequence-structure into a unified modality. We apply the created protein language on general backbone inpainting task, building the first GPT-style model (FoldGPT) for sequence-structure co-generation with promising results. Key to our success is the substantial enhancement of the vector quantization module, Soft Conditional Vector Quantization (SoftCVQ).Published
2025-04-11
How to Cite
Gao, Z., Tan, C., Wang, J., Huang, Y., Wu, L., & Li, S. Z. (2025). FoldToken: Learning Protein Language via Vector Quantization and Beyond. Proceedings of the AAAI Conference on Artificial Intelligence, 39(1), 219–227. https://doi.org/10.1609/aaai.v39i1.31998
Issue
Section
AAAI Technical Track on Application Domains