MK2 at PBIG Competition: A Prompt Generation Solution

📅 2025-07-11
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This study addresses the challenge of efficiently transforming patents into product ideation. We propose MK2, a lightweight, training-free prompt evolution framework. Methodologically, Gemini 2.5 iteratively refines an initial prompt by integrating effective fragments from low-quality outputs; GPT-4.1 generates product concepts using the optimized prompt; and Qwen3-8B orchestrates an Elo-based tournament to evaluate and select the best-performing prompts. Our key contribution lies in establishing a cross-model collaborative and game-theoretic prompt evolution mechanism, substantially enhancing the practicality of prompt engineering for commercial ideation. Experiments span three technical domains and employ six automated evaluation dimensions. MK2 outperforms baselines in 25 out of 36 comparative trials and achieves top overall performance—slightly trailing only in materials chemistry. Results confirm the framework’s strong generalizability and real-world applicability.

Technology Category

Application Category

📝 Abstract
The Patent-Based Idea Generation task asks systems to turn real patents into product ideas viable within three years. We propose MK2, a prompt-centric pipeline: Gemini 2.5 drafts and iteratively edits a prompt, grafting useful fragments from weaker outputs; GPT-4.1 then uses this prompt to create one idea per patent, and an Elo loop judged by Qwen3-8B selects the best prompt-all without extra training data. Across three domains, two evaluator types, and six criteria, MK2 topped the automatic leaderboard and won 25 of 36 tests. Only the materials-chemistry track lagged, indicating the need for deeper domain grounding; yet, the results show that lightweight prompt engineering has already delivered competitive, commercially relevant ideation from patents.
Problem

Research questions and friction points this paper is trying to address.

Generating viable product ideas from patents
Optimizing prompt engineering without training data
Evaluating idea quality across multiple domains
Innovation

Methods, ideas, or system contributions that make the work stand out.

Gemini 2.5 drafts and edits prompts iteratively
GPT-4.1 generates one idea per patent
Elo loop selects best prompt via Qwen3-8B
🔎 Similar Papers
No similar papers found.
Y
Yuzheng Xu
OMRON SINIC X, NexaScience
Tosho Hirasawa
Tosho Hirasawa
Omron Sinic X
Natural Language ProcessingMultimodal LearningMachine Translation
S
Seiya Kawano
Kyoto Institute of Technology, NexaScience
S
Shota Kato
Kyoto University
Tadashi Kozuno
Tadashi Kozuno
OMRON SINIC X
reinforcement learningmachine learningneuroscience