JanusDDG: A Thermodynamics-Compliant Model for Sequence-Based Protein Stability via Two-Fronts Multi-Head Attention

📅 2025-04-04
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the problem of accurately predicting the change in protein stability (ΔΔG) induced by single- and multi-point mutations from sequence-only input, while strictly enforcing fundamental thermodynamic constraints—antisymmetry and transitivity. We propose the first thermodynamically constrained bidirectional cross-attention mechanism: leveraging protein language model (PLM) embeddings as input, the query and value vectors encode the wild-type–mutant representation difference, while the key vector alternates to jointly enable perturbation-awareness and contextual preservation. Additionally, we introduce a physics-informed regularization term that explicitly enforces thermodynamically consistent ΔΔG relationships. Our method achieves state-of-the-art performance across multiple benchmark datasets, matching or surpassing structure-based approaches in accuracy—despite using no 3D structural information. To our knowledge, this is the first method enabling high-fidelity, structure-free prediction of multi-mutation ΔΔG while satisfying rigorous thermodynamic principles.

Technology Category

Application Category

📝 Abstract
Understanding how residue variations affect protein stability is crucial for designing functional proteins and deciphering the molecular mechanisms underlying disease-related mutations. Recent advances in protein language models (PLMs) have revolutionized computational protein analysis, enabling, among other things, more accurate predictions of mutational effects. In this work, we introduce JanusDDG, a deep learning framework that leverages PLM-derived embeddings and a bidirectional cross-attention transformer architecture to predict $Delta Delta G$ of single and multiple-residue mutations while simultaneously being constrained to respect fundamental thermodynamic properties, such as antisymmetry and transitivity. Unlike conventional self-attention, JanusDDG computes queries (Q) and values (V) as the difference between wild-type and mutant embeddings, while keys (K) alternate between the two. This cross-interleaved attention mechanism enables the model to capture mutation-induced perturbations while preserving essential contextual information. Experimental results show that JanusDDG achieves state-of-the-art performance in predicting $Delta Delta G$ from sequence alone, matching or exceeding the accuracy of structure-based methods for both single and multiple mutations.
Problem

Research questions and friction points this paper is trying to address.

Predicts protein stability changes from residue mutations
Ensures thermodynamic consistency in stability predictions
Uses transformer architecture for mutation effect analysis
Innovation

Methods, ideas, or system contributions that make the work stand out.

Bidirectional cross-attention transformer architecture
PLM-derived embeddings for stability prediction
Cross-interleaved attention mechanism
🔎 Similar Papers
No similar papers found.
G
Guido Barducci
Computational Biomedicine Unit, Dept. of Medical Sciences, University of Turin, Turin, Italy
I
I. Rossi
Computational Biomedicine Unit, Dept. of Medical Sciences, University of Turin, Turin, Italy
F
Francesco Codicè
Computational Biomedicine Unit, Dept. of Medical Sciences, University of Turin, Turin, Italy
Cesare Rollo
Cesare Rollo
Post-doctoral Researcher, Department of Medical Sciences, University of Turin
Deep LearningSurvival AnalysisFederated Learning
V
Valeria Repetto
Computational Biomedicine Unit, Dept. of Medical Sciences, University of Turin, Turin, Italy
Corrado Pancotti
Corrado Pancotti
AI Consultant in Healthcare at Helmholtz Munich
Machine learningPhysicsBiology
V
Virginia Iannibelli
Computational Biomedicine Unit, Dept. of Medical Sciences, University of Turin, Turin, Italy
T
T. Sanavia
Computational Biomedicine Unit, Dept. of Medical Sciences, University of Turin, Turin, Italy
Piero Fariselli
Piero Fariselli
Dept. of Medical Sciences, University of Torino, Italy
BioinformaticsComputational BiophysicsMachine LearningNature Photography