🤖 AI Summary
This work addresses multidimensional social biases in large language models (LLMs) stemming from training data. We introduce the first unified bias evaluation framework covering diverse sensitive attributes—including physical characteristics and socioeconomic factors. To enable systematic assessment, we propose five generalizable prompting strategies for automated, cross-bias-type detection; design implicit bias metrics—Stereotype Score and Bias Amplification Ratio—and establish a multi-benchmark comparative evaluation framework. Empirical analysis across state-of-the-art models reveals that all exhibit statistically significant bias in at least one dimension, with LLaMA3.1-8B showing the lowest overall bias severity. Crucially, we identify data contamination and instruction-alignment mismatch as primary root causes. Our study delivers a rigorous, reproducible methodology and an open empirical benchmark for LLM bias evaluation and mitigation.
📝 Abstract
Advancements in Large Language Models (LLMs) have increased the performance of different natural language understanding as well as generation tasks. Although LLMs have breached the state-of-the-art performance in various tasks, they often reflect different forms of bias present in the training data. In the light of this perceived limitation, we provide a unified evaluation of benchmarks using a set of representative LLMs that cover different forms of biases starting from physical characteristics to socio-economic categories. Moreover, we propose five prompting approaches to carry out the bias detection task across different aspects of bias. Further, we formulate three research questions to gain valuable insight in detecting biases in LLMs using different approaches and evaluation metrics across benchmarks. The results indicate that each of the selected LLMs suffer from one or the other form of bias with the LLaMA3.1-8B model being the least biased. Finally, we conclude the paper with the identification of key challenges and possible future directions.