Papers
arxiv:2508.04664

Sculptor: Empowering LLMs with Cognitive Agency via Active Context Management

Published on Aug 6
· Submitted by Mor-Li on Aug 7
Authors:
Mo Li ,
,
,
,

Abstract

Sculptor, a framework for Active Context Management, enhances LLM performance on long contexts by enabling proactive attention and memory control, reducing proactive interference and improving reasoning reliability.

AI-generated summary

Large Language Models (LLMs) suffer from significant performance degradation when processing long contexts due to proactive interference, where irrelevant information in earlier parts of the context disrupts reasoning and memory recall. While most research focuses on external memory systems to augment LLMs' capabilities, we propose a complementary approach: empowering LLMs with Active Context Management (ACM) tools to actively sculpt their internal working memory. We introduce Sculptor, a framework that equips LLMs with three categories of tools: (1) context fragmentation, (2) summary, hide, and restore, and (3) intelligent search. Our approach enables LLMs to proactively manage their attention and working memory, analogous to how humans selectively focus on relevant information while filtering out distractions. Experimental evaluation on information-sparse benchmarks-PI-LLM (proactive interference) and NeedleBench Multi-Needle Reasoning-demonstrates that Sculptor significantly improves performance even without specific training, leveraging LLMs' inherent tool calling generalization capabilities. By enabling Active Context Management, Sculptor not only mitigates proactive interference but also provides a cognitive foundation for more reliable reasoning across diverse long-context tasks-highlighting that explicit context-control strategies, rather than merely larger token windows, are key to robustness at scale.

Community

Paper author Paper submitter
edited 2 days ago

Our work addresses the challenges LLMs face with long contexts, where simply enlarging the context window isn't a complete solution and can lead to performance issues like position bias, information overload, and general interference. We focus on how to mitigate these problems, particularly 'proactive interference,' where older, irrelevant information disrupts current reasoning. To address this, we introduce the Sculptor framework, based on an idea we call Active Context Management (ACM). Instead of having the model passively process all information, this framework provides it with a toolkit to actively 'sculpt' and manage its own working memory—for example, by folding irrelevant passages to reduce noise or performing a quick search when needed. Our initial experiments on benchmarks like PI-LLM and NeedleBench demonstrate the effectiveness of this approach, suggesting that empowering models with explicit context-control abilities is a promising, complementary direction toward more reliable long-context reasoning.

This is an automated message from the Librarian Bot. I found the following papers similar to this paper.

The following papers were recommended by the Semantic Scholar API

Please give a thumbs up to this comment if you found it helpful!

If you want recommendations for any Paper on Hugging Face checkout this Space

You can directly ask Librarian Bot for paper recommendations by tagging it in a comment: @librarian-bot recommend

Sign up or log in to comment

Models citing this paper 0

No model linking this paper

Cite arxiv.org/abs/2508.04664 in a model README.md to link it from this page.

Datasets citing this paper 0

No dataset linking this paper

Cite arxiv.org/abs/2508.04664 in a dataset README.md to link it from this page.

Spaces citing this paper 0

No Space linking this paper

Cite arxiv.org/abs/2508.04664 in a Space README.md to link it from this page.

Collections including this paper 1