SEO: Stochastic Experience Optimization for Large Language Models

๐Ÿ“… 2025-01-08
๐Ÿ“ˆ Citations: 0
โœจ Influential: 0
๐Ÿ“„ PDF
๐Ÿค– AI Summary
This work addresses the limited task-specific generalization and heavy reliance on parameter fine-tuning exhibited by large language models (LLMs). To this end, we propose a gradient-free natural language experience optimization method. Our approach introduces an iterative, model-specific self-optimization framework under frozen-parameter conditions; a novel stochastic validation mechanism to ensure the effectiveness and robustness of experience updates; and a model-agnostic experience evaluation and selection strategy enabling cross-distribution data transfer. Experiments across three representative tasks and three mainstream LLMs demonstrate consistent performance gains. Crucially, the optimized experiences generalize effectively to out-of-distribution data and significantly enhance zero-shot transfer capability to semantically similar tasksโ€”without any parameter updates.

Technology Category

Application Category

๐Ÿ“ Abstract
Large Language Models (LLMs) can benefit from useful experiences to improve their performance on specific tasks. However, finding helpful experiences for different LLMs is not obvious, since it is unclear what experiences suit specific LLMs. Previous studies intended to automatically find useful experiences using LLMs, while it is difficult to ensure the effectiveness of the obtained experience. In this paper, we propose Stochastic Experience Optimization (SEO), an iterative approach that finds optimized model-specific experience without modifying model parameters through experience update in natural language. In SEO, we propose a stochastic validation method to ensure the update direction of experience, avoiding unavailing updates. Experimental results on three tasks for three LLMs demonstrate that experiences optimized by SEO can achieve consistently improved performance. Further analysis indicates that SEO-optimized experience can generalize to out-of-distribution data, boosting the performance of LLMs on similar tasks.
Problem

Research questions and friction points this paper is trying to address.

Large Language Models
Performance Enhancement
Transfer Learning
Innovation

Methods, ideas, or system contributions that make the work stand out.

Randomized Empirical Optimization (SEO)
Natural Language Learning
Generalization Improvement
๐Ÿ”Ž Similar Papers
No similar papers found.
Jitao Xu
Jitao Xu
PhD student, LIMSI
Machine Translation
Hongyun Zhou
Hongyun Zhou
Harbin Institute of Technology Master
PEFTMachine translationLLM
L
Lei Shen
NetEase Youdao, Beijing, China
C
Conghui Zhu
Faculty of Computing, Harbin Institute of Technology
J
Jin Huang
NetEase Youdao, Beijing, China
Yitao Duan
Yitao Duan
NetEase Youdao
Large scale machine learningdistributed systemscryptography