Have We Designed Generalizable Structural Knowledge Promptings? Systematic Evaluation and Rethinking

📅 2024-12-31
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Structured Knowledge Prompting (SKP) methods lack systematic evaluation of generalization capabilities across varying task difficulty, knowledge granularity, and domain shifts. Method: We propose a multidimensional generalizability analysis framework—assessing granularity sensitivity, cross-task transferability, scale scalability, and domain universality—and introduce SUBARU, the first benchmark supporting multi-granularity, multi-level evaluation across nine diverse tasks. We further design a standardized evaluation pipeline integrating structured knowledge injection, fine-grained modeling, and transfer validation. Results: Empirical analysis reveals severe generalization limitations in current SKP approaches: performance degrades significantly with finer-grained knowledge specifications and cross-task adaptation. Our findings provide rigorous empirical evidence for robust knowledge injection and identify critical directions for improvement—particularly in granularity-aware prompting, transfer-efficient architectures, and domain-adaptive knowledge grounding.

Technology Category

Application Category

📝 Abstract
Large language models (LLMs) have demonstrated exceptional performance in text generation within current NLP research. However, the lack of factual accuracy is still a dark cloud hanging over the LLM skyscraper. Structural knowledge prompting (SKP) is a prominent paradigm to integrate external knowledge into LLMs by incorporating structural representations, achieving state-of-the-art results in many knowledge-intensive tasks. However, existing methods often focus on specific problems, lacking a comprehensive exploration of the generalization and capability boundaries of SKP. This paper aims to evaluate and rethink the generalization capability of the SKP paradigm from four perspectives including Granularity, Transferability, Scalability, and Universality. To provide a thorough evaluation, we introduce a novel multi-granular, multi-level benchmark called SUBARU, consisting of 9 different tasks with varying levels of granularity and difficulty.
Problem

Research questions and friction points this paper is trying to address.

Structural Knowledge Guidance
Large Language Models
Task Performance
Innovation

Methods, ideas, or system contributions that make the work stand out.

SUBARU Testing Framework
Structural Knowledge Guidance
Large Language Models
Y
Yichi Zhang
Zhejiang University - Ant Group Joint Laboratory of Knowledge Graph, Zhejiang University
Z
Zhuo Chen
Zhejiang University - Ant Group Joint Laboratory of Knowledge Graph, Zhejiang University
Lingbing Guo
Lingbing Guo
Tianjin University
Machine learningArtificial Intelligence
Y
Yajing Xu
Zhejiang University - Ant Group Joint Laboratory of Knowledge Graph, Zhejiang University
S
Shaokai Chen
Zhejiang University - Ant Group Joint Laboratory of Knowledge Graph, Zhejiang University
Mengshu Sun
Mengshu Sun
Beijing University of Technology
Deep LearningModel Compression and Acceleration
Binbin Hu
Binbin Hu
BUPT & Ant Group
Deep LearningData MiningGraph EmbeddingRecommender System
Z
Zhiqiang Zhang
Ant Group
Lei Liang
Lei Liang
Ant Group
Knowledge GraphAI
W
Wen Zhang
Zhejiang University - Ant Group Joint Laboratory of Knowledge Graph, Zhejiang University
H
Huajun Chen
Zhejiang University - Ant Group Joint Laboratory of Knowledge Graph, Zhejiang University