Search papers, labs, and topics across Lattice.
8
0
9
0
Forget training separate models for each pedestrian attribute dataset – a single Transformer can now handle RGB images, video sequences, and even event streams with comparable accuracy to specialized methods.
Social robots can now autonomously orchestrate complex tasks with improved efficiency and emotional alignment, thanks to a novel fast-slow thinking LLM framework.
GTokenLLMs suffer from a text-dominant bias, but RGLM offers a way to fix this by reconstructing graph information directly from the LLM's graph token outputs.
Achieve state-of-the-art low-field MRI enhancement by explicitly modeling and correcting the intensity distribution shift between low-field and high-field domains using a differentiable Sinkhorn optimal transport module within a diffusion framework.
Event cameras can significantly boost the robustness of pre-trained OCR models for kilometer marker recognition in challenging metro environments, even under GNSS-denied conditions.
A nested Mixture-of-Experts architecture lets neural operators pre-trained on diverse PDEs transfer more effectively to downstream tasks.
A single tokenizer, UniWeTok, now handles both high-fidelity image reconstruction and complex semantic understanding for multimodal LLMs, outperforming existing methods with far less training data.
Unlock SOTA performance in long-horizon search tasks with REDSearcher, a framework that slashes the cost of training by strategically synthesizing complex tasks and boosting core LLM capabilities *before* reinforcement learning.