๐ค AI Summary
Existing short-context, multiple-choice bias benchmarks fail to assess free-text bias in large language model (LLM) deployments because they neglect prompt-scenario interactions, while fully manual evaluation is prohibitively costly. Method: This paper proposes a scalable human evaluation framework tailored to real-world deployment settings. It features (1) an operationalized, fine-grained taxonomy of bias types enabling free-text annotation; (2) a human-in-the-loop semi-automated pipeline integrating qualitative analysisโdriven rule modeling and bias pattern induction; and (3) systematic diagnosis exposing structural flaws in mainstream benchmark templates. Results: Evaluated across multiple LLM outputs, the framework uncovers context-sensitive biases entirely masked by conventional benchmarks, thereby substantially enhancing the authenticity and ecological validity of bias assessment.
๐ Abstract
LLM evaluation is challenging even the case of base models. In real world deployments, evaluation is further complicated by the interplay of task specific prompts and experiential context. At scale, bias evaluation is often based on short context, fixed choice benchmarks that can be rapidly evaluated, however, these can lose validity when the LLMs' deployed context differs. Large scale human evaluation is often seen as too intractable and costly. Here we present our journey towards developing a semi-automated bias evaluation framework for free text responses that has human insights at its core. We discuss how we developed an operational definition of bias that helped us automate our pipeline and a methodology for classifying bias beyond multiple choice. We additionally comment on how human evaluation helped us uncover problematic templates in a bias benchmark.