🤖 AI Summary
In the LLM era, programming assessment faces a fundamental tension between AI-assisted question generation and robust anti-cheating guarantees. Method: This paper proposes and implements an open-source, locally deployable technical assessment platform integrating LLM-driven problem generation, online code execution, time-constrained sandboxed evaluation, and an interactive web frontend—establishing an end-to-end closed-loop assessment pipeline. Contribution/Results: It is the first system enabling problem authors to autonomously design, publish, and automatically grade programming tasks entirely in offline or private environments, thereby balancing question diversity with assessment integrity. Its lightweight architecture ensures zero-cost deployment and high customizability, making it suitable for resource-constrained educational and recruitment settings. Experimental evaluation demonstrates significant improvements in assessment preparation efficiency and operational convenience, while preserving the validity and reliability of candidate ability measurement.
📝 Abstract
Organizations and educational institutions use time-bound assessment tasks to evaluate coding and problem-solving skills. These assessments measure not only the correctness of the solutions, but also their efficiency. Problem setters (educator/interviewer) are responsible for crafting these challenges, carefully balancing difficulty and relevance to create meaningful evaluation experiences. Conversely, problem solvers (student/interviewee) apply coding efficiency and logical thinking to arrive at correct solutions. In the era of Large Language Models (LLMs), LLMs assist problem setters in generating diverse and challenging questions, but they can undermine assessment integrity for problem solvers by providing easy access to solutions. This paper introduces OpenCoderRank, an easy-to-use platform designed to simulate technical assessments. It acts as a bridge between problem setters and problem solvers, helping solvers prepare for time constraints and unfamiliar problems while allowing setters to self-host assessments, offering a no-cost and customizable solution for technical assessments in resource-constrained environments.