🤖 AI Summary
Standard NLP models struggle to identify implicit cultural capital (CC) themes—such as aspirational goals and familial support—in STEM education narratives, primarily because they process sentences in isolation and lack domain-specific contextual understanding; these themes often manifest across sentence boundaries rather than through explicit keywords. To address this, we propose a “triple-aware” framework: domain-aware (incorporating STEM-adapted vocabulary), context-aware (modeling document-level sentence dependencies), and overlap-aware (capturing multi-label co-occurrence structures), implemented as an end-to-end Transformer-based multi-label classifier. Evaluated on real-world student reflection data, our model achieves a 2.1-percentage-point improvement in Macro-F1 over the strongest baseline, with consistent gains across all CC themes and strong generalization. This work constitutes the first systematic modeling of the deep semantic structure of cultural capital in educational narratives, establishing an interpretable NLP paradigm for equitable education assessment.
📝 Abstract
Identifying cultural capital (CC) themes in student reflections can offer valuable insights that help foster equitable learning environments in classrooms. However, themes such as aspirational goals or family support are often woven into narratives, rather than appearing as direct keywords. This makes them difficult to detect for standard NLP models that process sentences in isolation. The core challenge stems from a lack of awareness, as standard models are pre-trained on general corpora, leaving them blind to the domain-specific language and narrative context inherent to the data. To address this, we introduce AWARE, a framework that systematically attempts to improve a transformer model's awareness for this nuanced task. AWARE has three core components: 1) Domain Awareness, adapting the model's vocabulary to the linguistic style of student reflections; 2) Context Awareness, generating sentence embeddings that are aware of the full essay context; and 3) Class Overlap Awareness, employing a multi-label strategy to recognize the coexistence of themes in a single sentence. Our results show that by making the model explicitly aware of the properties of the input, AWARE outperforms a strong baseline by 2.1 percentage points in Macro-F1 and shows considerable improvements across all themes. This work provides a robust and generalizable methodology for any text classification task in which meaning depends on the context of the narrative.