Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: Designing protein sequences that fold into a target three-dimensional
structure, known as the inverse folding problem, is central to protein
engineering but remains challenging due to the vast sequence space and the
importance of local structural constraints. Existing deep learning approaches
achieve strong recovery rates, yet they lack explicit mechanisms to reuse
fine-grained structure-sequence patterns that are conserved across natural
proteins. We present PRISM, a multimodal retrieval-augmented generation
framework for inverse folding that retrieves fine-grained representations of
potential motifs from known proteins and integrates them with a hybrid
self-cross attention decoder. PRISM is formulated as a latent-variable
probabilistic model and implemented with an efficient approximation, combining
theoretical grounding with practical scalability. Across five benchmarks
(CATH-4.2, TS50, TS500, CAMEO 2022, and the PDB date split), PRISM establishes
new state of the art in both perplexity and amino acid recovery, while also
improving foldability metrics (RMSD, TM-score, pLDDT), demonstrating that
fine-grained multimodal retrieval is a powerful and efficient paradigm for
protein sequence design.
Authors (3)
Sazan Mahbub
Souvik Kundu
Eric P. Xing
Submitted
October 12, 2025
Key Contributions
Introduces PRISM, a retrieval-augmented generation framework for the inverse folding problem. It leverages fine-grained structure-sequence patterns from known proteins via retrieval and integrates them into a hybrid attention decoder. PRISM is formulated as a latent-variable probabilistic model, achieving new state-of-the-art results across multiple benchmarks by effectively reusing conserved motifs.
Business Value
Enables the rational design of novel proteins with specific functions, accelerating the development of new enzymes, therapeutics, biomaterials, and other protein-based technologies.