Search papers, labs, and topics across Lattice.

Top-tier US AI research university. Strong in NLP, ML systems, and computer vision.
45
93
2
A surprisingly simple tweak to Hartigan's k-means algorithm unlocks another 2-5% accuracy boost, especially when clustering high-dimensional data.
Kernel launch overhead is a bigger bottleneck than you think: GPUOS achieves up to 15.3x speedup by fusing operations at runtime.
Modular training with BAR allows independent updates of domain experts, achieving superior performance without the pitfalls of catastrophic forgetting.
RosettaSearch recovers up to 68% more structural fidelity in protein designs, transforming how we optimize sequences beyond traditional single-pass methods.
Geometric matrix interpolation reveals hidden common structures in multi-view data, offering a new lens for multi-manifold learning.
Massively multilingual NER just got easier: UNER v2 offers a standardized benchmark for evaluating LLMs across diverse languages.
LLMs are twice as likely as humans to repeat the same support tactic in a conversation, but a simple RL reward for tactic novelty can fix it.
Forget training on closed sets: WildDet3D leverages geometric cues and diverse prompts to achieve SOTA 3D object detection across 13.5K categories in the wild.
Achieving robust brain decoding across subjects without any retraining could revolutionize how we interpret neural signals in diverse populations.
MLLMs can be tricked into missing 90% of harmful content simply by encoding it in images that humans can easily read.
Get 80% of your prompt length back without sacrificing accuracy using a diffusion-based pruning method that can mask multiple tokens at once.
Serving both image and video diffusion models on the same hardware? GENSERVE's step-level preemption and dynamic resource allocation can boost your service level agreement (SLA) attainment by up to 44%.
Forget scaling laws: a large VLM strategically paired with a smaller model's reasoning tokens can rival the performance of a much larger, monolithic model.
Forget catastrophic forgetting: ProTPS leverages vision prototypes to guide text prompt learning, achieving near-upper-bound performance in continual learning scenarios.
Forget hand-designed agent communication topologies: Agent Q-Mix learns decentralized communication strategies that boost accuracy and token efficiency in LLM multi-agent systems.
Claims of quantum advantage in electronic structure calculations must now contend with DMRG benchmarks achieving CAS(89,102) on Fe$_5$S$_{12}$H$_4^{5-}$, pushing the boundaries of classical computation.
Generative multi-agent systems spontaneously exhibit collusion and conformity, mirroring societal pathologies, even without explicit programming and bypassing individual agent safeguards.
Today's best MLLMs are stumped by PerceptionComp, a new video reasoning benchmark where answering questions requires piecing together visual evidence across time and space.
AI can now (almost) write and direct Saturday Night Live.
Agentic search gets a meta-RL boost: MR-Search learns to self-reflect and adapt search strategies across episodes, significantly outperforming standard RL baselines.
AI interventions designed to combat ableism can backfire, as biased nudges were often rejected and increased negativity, while inclusive nudges proved more effective as scaffolding for learning.
LLM-powered VR guides for blind and low vision users are not just tools, but social actors, prompting users to give them nicknames and rationalize their mistakes when others are present.
See in the dark: Dark3R unlocks structure from motion at signal-to-noise ratios below -4dB, where existing methods completely break down.
Existing AI agent permissioning schemes are hard to compare, so this paper provides a formal foundation and reveals a fundamental conflict between training data confidentiality and agent completeness.
Learning robotic reward functions from a million trajectories reveals that comparing entire trajectories, not just individual frames, unlocks better generalization and learning from suboptimal data.
LLMs still struggle with factual accuracy in specialized medical domains like pancreatic cancer, with hallucination rates varying wildly and web search integration failing to guarantee better responses.
Forget full fine-tuning: this dynamic routing strategy lets you adapt dense retrieval to new domains while using just 2% of the parameters.
Hyperspectral video, previously limited by motion artifacts and poor photon utilization, now achieves real-time capture and improved fidelity thanks to active illumination and coded-exposure pixels.
No-regret learning in repeated Bertrand games can lead to surprisingly high prices, challenging classical game theory's low-price predictions.
Unlock robot learning with hidden knowledge: TOPReward extracts surprisingly accurate task progress signals directly from VLM token probabilities, bypassing the need for explicit reward engineering.
Forget passively analyzing model outputs – this new attack actively *trains* the model to regurgitate specific texts, revealing its training data with surprising accuracy.
Key contribution not extracted.
LMs can learn some human-like linguistic biases from synthetic data, but surprisingly fail to reproduce the strong object preference seen in differential argument marking across human languages.
Stop worrying about false positives: this watermarking scheme guarantees unforgeability and recoverability, ensuring content is linked exclusively to its generating model even under substitution attacks.
Forget RL fine-tuning: this paper shows you can beat it at cold-start personalization with a tiny model and clever Bayesian inference over structured preference priors.
Forget synthetic benchmarks that don't translate: MolmoSpaces offers 230k diverse, simulator-agnostic environments with 130k annotated objects, showing a remarkable 0.96 sim-to-real correlation for robot policies.
Open-weight coding agents can now be cheaply and rapidly specialized to private codebases, thanks to a new supervised finetuning method that slashes training costs by over 25x.
This study establishes SSL as a promising paradigm for ECG analysis, particularly in settings with limited annotated data, enhancing accessibility, generalizability, and fairness in AI-driven cardiac diagnostics across diverse clinical environments and questions.
Moxin 7B and its variants (VLM, VLA, Chinese) offer a new suite of fully transparent, open-source multimodal models, pushing beyond simple weight sharing to enable deeper customization and collaborative research.
Robots can now navigate more reliably and across different bodies (wheeled vs. legged) thanks to a hierarchical model that separates high-level planning from low-level physical constraints.
Open-source biomolecular modeling just got a boost: RF3 closes the gap with AlphaFold3 in structure prediction, thanks to the new AtomWorks data framework.
Robot foundation models can achieve state-of-the-art performance by explicitly reasoning about spatial plans as editable trajectory traces, rather than directly mapping perception to control.
Train better aligned LLMs with 10% of the data by strategically focusing on the most difficult preference comparisons.
Despite claims of safety alignment, state-of-the-art LLMs still spill the beans on hazardous scientific knowledge at an alarming rate, failing nearly 80% of the time on a new regulation-grounded benchmark.
Self-supervised learning beats supervised learning for ECG interpretation when labeled data is scarce, unlocking more robust and generalizable AI-driven cardiac diagnostics.