🤖 AI Summary
This paper addresses the growing safety and ethical risks—such as bias, harmful content, and cultural insensitivity—posed by large language models (LLMs) in children’s applications, compounded by the absence of standardized evaluation frameworks. Methodologically, it integrates a systematic literature review, risk taxonomy development, and cross-cultural empirical studies to construct a three-dimensional assessment framework encompassing toxicity detection, ethical alignment, and cultural adaptability, alongside operational evaluation tools and implementation guidelines. The primary contribution is the first child-centered LLM safety evaluation paradigm that explicitly incorporates parental concerns and diverse cultural contexts, thereby filling a critical gap in AI safety standardization for children. It provides evidence-based, actionable governance pathways for developers, educators, and policymakers engaged in responsible AI deployment for young users.
📝 Abstract
This study examines the growing use of Large Language Models (LLMs) in child-centered applications, highlighting safety and ethical concerns such as bias, harmful content, and cultural insensitivity. Despite their potential to enhance learning, there is a lack of standardized frameworks to mitigate these risks. Through a systematic literature review, we identify key parental and empirical concerns, including toxicity and ethical breaches in AI outputs. Moreover, to address these issues, this paper proposes a protection framework for safe Child-LLM interaction, incorporating metrics for content safety, behavioral ethics, and cultural sensitivity. The framework provides practical tools for evaluating LLM safety, offering guidance for developers, policymakers, and educators to ensure responsible AI deployment for children.