🤖 AI Summary
To address the inconsistent quality of online educational videos and the lack of automated assessment tools, this paper introduces the first large-scale multimodal dataset of instructional videos paired with corresponding teaching materials. We systematically formulate three foundational principles for high-quality educational videos—pedagogical soundness, structural coherence, and expressive clarity—and innovatively cast video quality assessment as a structured multiple-choice question-answering task. Our method integrates pedagogy-aware prompt engineering, language-model-driven fine-grained scoring, and principle-based interpretability analysis. Evaluated on our newly constructed benchmark, our approach achieves a 23.5% accuracy improvement over baseline methods and effectively discriminates between high- and low-quality courses. Crucially, scores are fully attributable and traceable to specific pedagogical principles, enabling actionable, instructor-targeted feedback and data-informed course refinement.
📝 Abstract
Online courses have significantly lowered the barrier to accessing education, yet the varying content quality of these videos poses challenges. In this work, we focus on the task of automatically evaluating the quality of video course content. We have constructed a dataset with a substantial collection of video courses and teaching materials. We propose three evaluation principles and design a new evaluation framework, extit{VCEval}, based on these principles. The task is modeled as a multiple-choice question-answering task, with a language model serving as the evaluator. Our method effectively distinguishes video courses of different content quality and produces a range of interpretable results.