🤖 AI Summary
Current approaches to assessing the novelty of scientific papers rely heavily on subjective judgment and lack systematic, interpretable, and reviewer-aligned objective methods. This work proposes the first knowledge-driven framework for novelty evaluation, explicitly modeling human judgments of novelty derived from peer review comments across nearly 80,000 top-tier AI conference papers. By integrating structured paper representations with a semantic similarity graph of related literature, the framework enables fine-grained, concept-level originality comparisons. It combines large language model fine-tuning, knowledge extraction, and semantic retrieval to produce calibrated, interpretable novelty scores that significantly outperform existing methods in accuracy, consistency, and alignment with human reviewers.
📝 Abstract
Assessing research novelty is a core yet highly subjective aspect of peer review, typically based on implicit judgment and incomplete comparison to prior work. We introduce a literature-aware novelty assessment framework that explicitly learns how humans judge novelty from peer-review reports and grounds these judgments in structured comparison to existing research. Using nearly 80K novelty-annotated reviews from top-tier AI conferences, we fine-tune a large language model to capture reviewer-aligned novelty evaluation behavior. For a given manuscript, the system extracts structured representations of its ideas, methods, and claims, retrieves semantically related papers, and constructs a similarity graph that enables fine-grained, concept-level comparison to prior work. Conditioning on this structured evidence, the model produces calibrated novelty scores and human-like explanatory assessments, reducing overestimation and improving consistency relative to existing approaches.