How Fair is Software Fairness Testing?

πŸ“… 2026-03-12
πŸ“ˆ Citations: 0
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
This study addresses the prevalent assumption in software fairness testing that fairness is a universal and context-independent construct, which often marginalizes non-Western epistemologies and vulnerable populations. Adopting an interdisciplinary critical lens grounded in cultural perspectives, the research systematically examines cultural biases embedded in fairness metrics, testing datasets, and ethical practices. It uniquely integrates insights from oral traditions, Indigenous languages, and non-digital communities to argue for moving beyond one-size-fits-all fairness indicators. The work proposes a culturally pluralistic and context-sensitive framework for evaluating algorithmic fairness, exposing systemic shortcomings in current approaches regarding cultural representativeness, data justice, and environmental ethics. By doing so, it lays a theoretical foundation and charts an innovative pathway toward more inclusive AI evaluation systems.

Technology Category

Application Category

πŸ“ Abstract
Software fairness testing is a central method for evaluating AI systems, yet the meaning of fairness is often treated as fixed and universally applicable. This vision paper positions fairness testing as culturally situated and examines the problem across three dimensions. First, fairness metrics encode particular cultural values while marginalizing others. Second, test datasets are predominantly designed from Western contexts, excluding knowledge systems grounded in oral traditions, Indigenous languages, and non-digital communities. Third, fairness testing raises ethical concerns, including the reliance on low-paid data labeling in the Global South, and associated with this, the environmental costs of training and deploying large-scale models, which disproportionately affect climate-vulnerable populations. Addressing these issues requires rethinking fairness testing beyond universal metrics and moving toward evaluation frameworks that respect cultural plurality and acknowledge the right to refuse algorithmic mediation.
Problem

Research questions and friction points this paper is trying to address.

software fairness testing
cultural situatedness
fairness metrics
test datasets
ethical concerns
Innovation

Methods, ideas, or system contributions that make the work stand out.

culturally situated fairness
fairness testing
algorithmic justice
decolonial AI
ethical AI evaluation