Search papers, labs, and topics across Lattice.
The paper introduces SLATE, a reinforcement learning framework for training LLMs to reason with search engines, addressing the credit assignment problem by using truncated step-level sampling and dense LLM-as-judge rewards. Truncated sampling reduces gradient variance by generating trajectories that share a common prefix, while LLM-as-judge rewards replace heuristic scoring with a more reliable LLM evaluator. Experiments across seven QA benchmarks demonstrate that SLATE outperforms sparse-reward and process-reward baselines, particularly on multi-hop tasks and with smaller models.
Forget sparse rewards: SLATE uses LLMs to judge each reasoning step, slashing gradient variance and boosting performance on retrieval-augmented reasoning tasks, especially for smaller models.
Training large language models to reason with search engines via reinforcement learning is hindered by a fundamental credit assignment problem: existing methods such as Search-R1 provide only a sparse outcome reward after an entire multi-step trajectory, making it infeasible to attribute success or failure to individual reasoning and retrieval decisions. Process-reward methods like StepSearch alleviate this by introducing step-level supervision, but rely on heuristic rewards such as TF-IDF overlap with gold documents, and still sample k complete trajectories per example, retaining high gradient variance. We propose SLATE, a framework built on two complementary ideas: (1) truncated step-level sampling, which generates k trajectories that share a common prefix and differ only at the next step, and (2) dense LLM-as-judge rewards, which replace heuristic scoring with a capable LLM evaluator that assesses the quality of each reasoning step, search query, and answer, providing richer and more reliable supervision. We theoretically prove that under the same dense reward structure, truncated sampling reduces the variance of advantage estimates by up to a factor of T compared to full-trajectory sampling for T-step trajectories, yielding lower-variance, better-targeted policy gradients. Experiments on seven QA benchmarks confirm that SLATE consistently outperforms both sparse-reward and process-reward baselines, with the largest gains on harder multi-hop tasks and smaller models.