AutoMLGen: Navigating Fine-Grained Optimization for Coding Agents

๐Ÿ“… 2025-10-09
๐Ÿ“ˆ Citations: 0
โœจ Influential: 0
๐Ÿ“„ PDF
๐Ÿค– AI Summary
Automated code optimization in machine learning engineering (MLE) suffers from heavy reliance on expert intervention, low search efficiency, and insufficient cross-path knowledge reuse. Method: This paper proposes AutoMLGen, an LLM-driven coding agent that integrates a domain-specific knowledge base with Monte Carlo Graph Search (MCGS)โ€”a graph-structured enhancement over conventional tree- or sequence-based search. MCGS enables dynamic path recombination, historical trajectory reuse, and multi-solution fusion to improve self-evolution capability and search diversity; combined with a fine-grained operator set and iterative learning, it further enhances stability and convergence speed. Results: On the MLE-Bench benchmark, AutoMLGen achieves state-of-the-art (SOTA) performance within 12 hoursโ€”50% of the standard budgetโ€”with significantly higher average medal rate and effective submission rate across multiple metrics.

Technology Category

Application Category

๐Ÿ“ Abstract
Large language models (LLMs) have shown impressive performance in general programming tasks. However, in Machine Learning Engineering (MLE) scenarios such as AutoML and Kaggle competitions, achieving high performance depends heavily on expert intervention and repeated adjustments rather than simply generating correct code. When applied directly to these tasks, LLMs often lack fine-grained domain priors, and existing MLE approaches that use linear or tree-structured searches limit knowledge transfer to adjacent hierarchical links. As a result, they cannot leverage past full trajectories or share information across branches, limiting self-evolving ability and search space diversity. To address these limitations, we introduce AutoMLGen, an LLM-based coding agent that integrates a domain knowledge base for high-quality prior guidance and Monte Carlo Graph Search (MCGS) for efficient exploration. MCGS retains the tree-guided exploration of MCTS while embedding a graph structure into the expansion stage to enable dynamic path reorganization, historical trajectory reuse, and multi-solution fusion to support both self-evolution and collaborative learning. Combined with fine-grained operator sets, this design improves stability and accelerates convergence. Evaluation on the MLE-Bench shows that AutoMLGen achieves state-of-the-art performance in numerous dimensions, such as the average medal rate and the valid submission rate, under a 12-hour budget (half the standard runtime). The code is available at https://github.com/Alpha-Innovator/InternAgent.
Problem

Research questions and friction points this paper is trying to address.

LLMs lack fine-grained domain knowledge for ML engineering tasks
Existing methods limit knowledge transfer and search space diversity
Need efficient exploration with historical trajectory reuse capabilities
Innovation

Methods, ideas, or system contributions that make the work stand out.

Integrates domain knowledge base for prior guidance
Uses Monte Carlo Graph Search for efficient exploration
Enables dynamic path reorganization and trajectory reuse
๐Ÿ”Ž Similar Papers
No similar papers found.
S
Shangheng Du
Shanghai Artificial Intelligence Laboratory, East China Normal University
Xiangchao Yan
Xiangchao Yan
Shanghai Jiao Tong University
Dengyang Jiang
Dengyang Jiang
Northwestern Polytechnical University
Computer VisionDeep LearningMachine Learning
Jiakang Yuan
Jiakang Yuan
Fudan university
MLLMsMulti-agent SystemReasoning
Y
Yusong Hu
Shanghai Artificial Intelligence Laboratory
X
Xin Li
Shanghai Artificial Intelligence Laboratory
L
Liang He
Shanghai Artificial Intelligence Laboratory, East China Normal University
B
Bo Zhang
Shanghai Artificial Intelligence Laboratory
Lei Bai
Lei Bai
Shanghai AI Laboratory
Foundation ModelScience IntelligenceMulti-Agent SystemAutonomous Discovery