Search papers, labs, and topics across Lattice.
Mimicking human cognition, FLAIR lets dialogue models "think while listening," boosting performance without adding latency.
Ditch the cross-world counterfactuals: Sequential Transport offers a DAG-aware, optimal transport approach to causal mediation analysis, providing deterministic counterfactual mediators and fine-grained attribution.
Forget brute-force scaling: Tiny Aya proves a 3B parameter model can achieve state-of-the-art multilingual performance with clever training and region-aware specialization.
Forget contrastive learning: LLM2Vec-Gen learns text embeddings by representing the *response* an LLM would generate, unlocking safety and reasoning abilities for embedding tasks.
One in four initial posts on a major cybercrime forum contain explicit crime-related content, revealing a surprisingly high baseline of open criminal activity.
Diagonal SSMs, despite their empirical success, provably fail to track states of non-Abelian groups, revealing fundamental limitations in their expressive power.
Forget full fine-tuning: this dynamic routing strategy lets you adapt dense retrieval to new domains while using just 2% of the parameters.
Achieve state-of-the-art dynamic graph anomaly detection with limited labels by learning a robust decision boundary around normal data, outperforming methods that overfit to scarce anomalies.
Attention-based re-ranking gets a boost: ReAttn's post-hoc re-weighting tames over-concentration and lexical bias, leading to more accurate and interpretable results without extra training.
LLMs struggle to balance rational financial decisions with mimicking noisy user behavior, often overfitting to short-term market trends instead of aligning with long-term investment goals.
Coreference benchmarks may be overstating language models' NLU abilities, as even small changes to evaluation contexts reveal a failure to generalize.
Cybercriminals are actively exploring AI's potential for both enhancing existing attacks and creating novel illicit tools, but harbor significant doubts about its real-world effectiveness and impact on their operations.
Dramatically improve protein language models by simply post-training them to align with protein graphs, yielding a 59% increase in contact prediction accuracy.
Forget retraining from scratch: port fine-tuning updates between LLM versions and get up to 47% performance boost on tasks like instruction following, even surpassing fully fine-tuned models.
Forget Bayesian bells and whistles: in-context learning shines brightest with simple point estimators, outperforming complex posterior approximations in most scenarios.