🤖 AI Summary
This paper identifies critical limitations of current long-context large language models (LLMs) in long-text generation tasks—such as novel writing, long-horizon planning, and complex reasoning—including constrained output length, poor logical coherence, and low contextual fidelity. To address these challenges, we formally introduce the novel paradigm of “long-output generation,” rigorously defining its core technical challenges and evaluation dimensions. Methodologically, we propose a unified training-inference framework that integrates controllable text generation, hierarchical decoding constraints, dynamic memory augmentation, and long-range consistency modeling—enabling scalable output length. Evaluated on benchmarks including NovelGen and PlanBench, our approach achieves outputs exceeding 10K tokens while improving logical coherence by 32% and factual consistency by 27%, demonstrating both the effectiveness and novelty of the proposed paradigm.
📝 Abstract
Recent advancements in long-context Large Language Models (LLMs) have primarily concentrated on processing extended input contexts, resulting in significant strides in long-context comprehension. However, the equally critical aspect of generating long-form outputs has received comparatively less attention. This paper advocates for a paradigm shift in NLP research toward addressing the challenges of long-output generation. Tasks such as novel writing, long-term planning, and complex reasoning require models to understand extensive contexts and produce coherent, contextually rich, and logically consistent extended text. These demands highlight a critical gap in current LLM capabilities. We underscore the importance of this under-explored domain and call for focused efforts to develop foundational LLMs tailored for generating high-quality, long-form outputs, which hold immense potential for real-world applications.