Search papers, labs, and topics across Lattice.
Xidian University
2
0
6
0
Forget fixed-precision quantization: STQuant slashes optimizer memory by 84% in large model training by dynamically adapting bit-widths across layers and training steps.
LLMs still fail to follow complex instructions that entangle content, formatting, control flow, and real-world constraints, despite progress on simpler benchmarks.