Search papers, labs, and topics across Lattice.
1
0
3
2
Sub-2-bit LLMs can now achieve state-of-the-art performance thanks to pQuant, which selectively preserves sensitive parameters in a high-precision branch during quantization-aware training.