Search papers, labs, and topics across Lattice.
The paper introduces Budget-Aware Context Management (BACM), a framework that formulates context management for LLM agents as a sequential decision problem under a context budget constraint. BACM-RL, an end-to-end RL approach, learns compression strategies to decide when and how much of the interaction history to compress based on the available budget. Experiments on multi-objective QA and web browsing show BACM-RL outperforms existing methods, achieving up to 1.6x gains in complex settings and maintaining performance as context budgets shrink.
LLM agents can navigate long-horizon tasks far more effectively by dynamically compressing their interaction history based on a learned context budget, outperforming fixed-strategy baselines by up to 60% in complex scenarios.
LLM-based agents show strong potential for long-horizon reasoning, yet their context size is limited by deployment factors (e.g., memory, latency, and cost), yielding a constrained context budget. As interaction histories grow, this induces a trade-off between retaining past information and staying within the context limit. To address this challenge, we propose Budget-Aware Context Management (BACM), which formulates context management as a sequential decision problem with a context budget constraint. It enables agents to assess the available budget before incorporating new observations and decide when and how much of the interaction history to compress. We further develop BACM-RL, an end-to-end curriculum-based reinforcement learning approach that learns compression strategies under varying context budgets. Experiments on compositional multi-objective QA and long-horizon web browsing benchmarks show that BACM-RL consistently outperforms prior methods across model scales and task complexities, achieving over $1.6\times$ gains over strong baselines in high-complexity settings, while maintaining strong advantages as budgets shrink, where most methods exhibit a downward performance trend.