🤖 AI Summary
Template-free retrosynthetic prediction suffers from low accuracy, poor robustness, and limited practical applicability in synthetic planning. Method: This paper proposes a graph-aware, template-free Transformer framework. It pioneers the direct encoding of molecular graph topology into the self-attention mechanism, enabling joint modeling of SMILES sequences and molecular graphs; introduces graph-enhanced attention and paired-data augmentation to improve generalization; and eliminates reaction templates and rule-based engines, enabling end-to-end generation of interpretable retrosynthetic pathways. Contribution/Results: It is the first work to achieve deep integration of graph-structural priors with the Transformer architecture, substantially enhancing modeling capability for complex reactions. On the USPTO-50K benchmark, it achieves state-of-the-art Top-1 accuracy among template-free methods, significantly outperforming the standard Transformer baseline.
📝 Abstract
Retrosynthesis reaction prediction seeks to infer plausible reactant molecules for a given product and is a central problem in computer-aided organic synthesis. Despite recent progress, many existing models still fall short of the accuracy and robustness required for practical deployment. This work studies a template-free, Transformer-based framework that eliminates reliance on handcrafted reaction templates or additional chemical rule engines. The model injects molecular graph information into the attention mechanism to jointly exploit SMILES sequences and structural cues, and further applies a paired data augmentation strategy to enhance training diversity and scale. On the USPTO-50K benchmark, our proposed approach achieves state-of-the-art performance among template-free methods and substantially outperforming a vanilla Transformer baseline.