🤖 AI Summary
Current recommendation system research automation focuses narrowly on isolated tasks—such as hyperparameter tuning—while neglecting end-to-end autonomy across the entire scientific workflow. To address this gap, we propose AutoRecLab: an “Autonomous Recommendation Systems Research Laboratory” paradigm that enables full-cycle automation—from problem identification and literature synthesis to experimental design, execution, result interpretation, paper drafting, and provenance tracking. Methodologically, AutoRecLab integrates large language model–driven problem generation, a multi-agent scientific discovery system, programmable experimental pipelines, and rigorous research provenance logging. Its core contributions are threefold: (1) the first formal definition and implementation of a fully AI-augmented, closed-loop research pipeline for recommender systems; (2) an open-source prototype platform, standardized benchmarking competition, AI-assisted paper review channel, and reproducibility framework; and (3) a foundational shift toward transparent, collaborative, and verifiable next-generation recommendation research.
📝 Abstract
Recommender-systems research has accelerated model and evaluation advances, yet largely neglects automating the research process itself. We argue for a shift from narrow AutoRecSys tools -- focused on algorithm selection and hyper-parameter tuning -- to an Autonomous Recommender-Systems Research Lab (AutoRecLab) that integrates end-to-end automation: problem ideation, literature analysis, experimental design and execution, result interpretation, manuscript drafting, and provenance logging. Drawing on recent progress in automated science (e.g., multi-agent AI Scientist and AI Co-Scientist systems), we outline an agenda for the RecSys community: (1) build open AutoRecLab prototypes that combine LLM-driven ideation and reporting with automated experimentation; (2) establish benchmarks and competitions that evaluate agents on producing reproducible RecSys findings with minimal human input; (3) create review venues for transparently AI-generated submissions; (4) define standards for attribution and reproducibility via detailed research logs and metadata; and (5) foster interdisciplinary dialogue on ethics, governance, privacy, and fairness in autonomous research. Advancing this agenda can increase research throughput, surface non-obvious insights, and position RecSys to contribute to emerging Artificial Research Intelligence. We conclude with a call to organise a community retreat to coordinate next steps and co-author guidance for the responsible integration of automated research systems.