Search papers, labs, and topics across Lattice.
Huazhong University of Sci. & Tech.
2
0
5
1
4
Achieve 97x speedup and 208x energy efficiency for Multi-Scale Deformable Attention by co-designing a near-memory processing architecture that overcomes the limitations of existing NMP solutions.
Moxin 7B and its variants (VLM, VLA, Chinese) offer a new suite of fully transparent, open-source multimodal models, pushing beyond simple weight sharing to enable deeper customization and collaborative research.