Abstract
A generative language model-based framework optimizes complex systems through stochastic generative optimization with prioritized exploration and contextual aggregation, achieving efficient convergence in both deterministic and stochastic environments.
Optimizing complex systems, ranging from LLM prompts to multi-turn agents, traditionally requires labor-intensive manual iteration. We formalize this challenge as a stochastic generative optimization problem where a generative language model acts as the optimizer, guided by numerical rewards and text feedback to discover the best system. We introduce Prioritized Optimization with Local Contextual Aggregation (POLCA), a scalable framework designed to handle stochasticity in optimization -- such as noisy feedback, sampling minibatches, and stochastic system behaviors -- while effectively managing the unconstrained expansion of solution space. POLCA maintains a priority queue to manage the exploration-exploitation tradeoff, systematically tracking candidate solutions and their evaluation histories. To enhance efficiency, we integrate an varepsilon-Net mechanism to maintain parameter diversity and an LLM Summarizer to perform meta-learning across historical trials. We theoretically prove that POLCA converges to near-optimal candidate solutions under stochasticity. We evaluate our framework on diverse benchmarks, including Ï„-bench, HotpotQA (agent optimization), VeriBench (code translation) and KernelBench (CUDA kernel generation). Experimental results demonstrate that POLCA achieves robust, sample and time-efficient performance, consistently outperforming state-of-the-art algorithms in both deterministic and stochastic problems. The codebase for this work is publicly available at https://github.com/rlx-lab/POLCA.
Community
We found that LLM-based optimization loop can benefit from Gemini embeddings -- we proposed an epsilon-net mechanism to accept/reject candidate proposals. This simple mechanism has theoretical guarantees and also performs well empirically, beating GEPA and OpenEvolve.
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- TSR: Trajectory-Search Rollouts for Multi-Turn RL of LLM Agents (2026)
- GFlowPO: Generative Flow Network as a Language Model Prompt Optimizer (2026)
- ContextEvolve: Multi-Agent Context Compression for Systems Code Optimization (2026)
- AdaEvolve: Adaptive LLM Driven Zeroth-Order Optimization (2026)
- MAGE: Meta-Reinforcement Learning for Language Agents toward Strategic Exploration and Exploitation (2026)
- Decoding ML Decision: An Agentic Reasoning Framework for Large-Scale Ranking System (2026)
- Bootstrapping Exploration with Group-Level Natural Language Feedback in Reinforcement Learning (2026)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment: @librarian-bot recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper
Collections including this paper 0
No Collection including this paper