🤖 AI Summary
This study rigorously evaluates whether proprietary large language models (LLMs) remain irreplaceable for automated essay scoring (AES), addressing critical concerns regarding performance, fairness, and cost-efficiency. Method: We systematically benchmark nine leading closed- and open-source LLMs—including GPT-4, Llama 3, and Qwen2.5—using few-shot prompting, semantic embedding analysis, ML-driven scoring regression, and disparate impact scoring to assess predictive accuracy, demographic fairness (by age and race), and generated essay quality. Contribution/Results: Empirical results show no statistically significant differences between top-tier open-source models (e.g., Llama 3, Qwen2.5) and GPT-4 in accuracy or fairness; moreover, their inference cost is only ~2.7% of GPT-4’s. This work provides the first empirical evidence that high-quality open-source LLMs can simultaneously achieve competitive accuracy, equitable outcomes, and operational affordability—establishing a foundational basis for scalable, inclusive deployment of AI in educational assessment.
📝 Abstract
Closed large language models (LLMs) such as GPT-4 have set state-of-the-art results across a number of NLP tasks and have become central to NLP and machine learning (ML)-driven solutions. Closed LLMs' performance and wide adoption has sparked considerable debate about their accessibility in terms of availability, cost, and transparency. In this study, we perform a rigorous comparative analysis of nine leading LLMs, spanning closed, open, and open-source LLM ecosystems, across text assessment and generation tasks related to automated essay scoring. Our findings reveal that for few-shot learning-based assessment of human generated essays, open LLMs such as Llama 3 and Qwen2.5 perform comparably to GPT-4 in terms of predictive performance, with no significant differences in disparate impact scores when considering age- or race-related fairness. Moreover, Llama 3 offers a substantial cost advantage, being up to 37 times more cost-efficient than GPT-4. For generative tasks, we find that essays generated by top open LLMs are comparable to closed LLMs in terms of their semantic composition/embeddings and ML assessed scores. Our findings challenge the dominance of closed LLMs and highlight the democratizing potential of open LLMs, suggesting they can effectively bridge accessibility divides while maintaining competitive performance and fairness.