Search papers, labs, and topics across Lattice.
4
0
8
8
Cut LLM cold starts from minutes to seconds by pre-materializing CUDA graph execution contexts, sidestepping brittle kernel patching and heavyweight checkpointing.
Scaling prompt learning by 17x without sacrificing accuracy is now possible, unlocking efficient self-improvement for LLM agents.
LLM-driven program evolution gets a smart upgrade: AdaEvolve dynamically allocates resources to promising solution candidates, leaving static schedules in the dust.
LLMs can now design GPU kernels that outperform both human experts and prior automated methods, thanks to a co-evolving world model that guides the search process.