Search papers, labs, and topics across Lattice.
4
0
7
4
By cleverly combining YOCO's efficient attention with recursive computation, YOCO-U achieves a capability-efficiency sweet spot that neither technique can reach on its own.
Language models can learn directly from real-world user interactions, boosting performance without human annotations or simulated environments.
1.58-bit LLMs are surprisingly more resilient to sparsity than their full-precision counterparts, opening new avenues for extreme compression.
Unlock 33% faster LLM inference on commodity GPUs with SlideSparse, which finally brings hardware-accelerated (2N-2):2N sparsity to the masses, bridging the accuracy gap left by NVIDIA's strict 2:4 pruning.