Sculptor: Empowering LLMs with Cognitive Agency via Active Context Management
Abstract
Sculptor, a framework for Active Context Management, enhances LLM performance on long contexts by enabling proactive attention and memory control, reducing proactive interference and improving reasoning reliability.
Large Language Models (LLMs) suffer from significant performance degradation when processing long contexts due to proactive interference, where irrelevant information in earlier parts of the context disrupts reasoning and memory recall. While most research focuses on external memory systems to augment LLMs' capabilities, we propose a complementary approach: empowering LLMs with Active Context Management (ACM) tools to actively sculpt their internal working memory. We introduce Sculptor, a framework that equips LLMs with three categories of tools: (1) context fragmentation, (2) summary, hide, and restore, and (3) intelligent search. Our approach enables LLMs to proactively manage their attention and working memory, analogous to how humans selectively focus on relevant information while filtering out distractions. Experimental evaluation on information-sparse benchmarks-PI-LLM (proactive interference) and NeedleBench Multi-Needle Reasoning-demonstrates that Sculptor significantly improves performance even without specific training, leveraging LLMs' inherent tool calling generalization capabilities. By enabling Active Context Management, Sculptor not only mitigates proactive interference but also provides a cognitive foundation for more reliable reasoning across diverse long-context tasks-highlighting that explicit context-control strategies, rather than merely larger token windows, are key to robustness at scale.
Community
Our work addresses the challenges LLMs face with long contexts, where simply enlarging the context window isn't a complete solution and can lead to performance issues like position bias, information overload, and general interference. We focus on how to mitigate these problems, particularly 'proactive interference,' where older, irrelevant information disrupts current reasoning. To address this, we introduce the Sculptor framework, based on an idea we call Active Context Management (ACM). Instead of having the model passively process all information, this framework provides it with a toolkit to actively 'sculpt' and manage its own working memory—for example, by folding irrelevant passages to reduce noise or performing a quick search when needed. Our initial experiments on benchmarks like PI-LLM and NeedleBench demonstrate the effectiveness of this approach, suggesting that empowering models with explicit context-control abilities is a promising, complementary direction toward more reliable long-context reasoning.
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- MemTool: Optimizing Short-Term Memory Management for Dynamic Tool Calling in LLM Agent Multi-Turn Conversations (2025)
- StoryBench: A Dynamic Benchmark for Evaluating Long-Term Memory with Multi Turns (2025)
- Beyond Isolated Capabilities: Bridging Long CoT Reasoning and Long-Context Understanding (2025)
- Evaluating Memory in LLM Agents via Incremental Multi-Turn Interactions (2025)
- MEM1: Learning to Synergize Memory and Reasoning for Efficient Long-Horizon Agents (2025)
- PaceLLM: Brain-Inspired Large Language Models for Long-Context Understanding (2025)
- World-aware Planning Narratives Enhance Large Vision-Language Model Planner (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper