Search papers, labs, and topics across Lattice.
3
9
6
4
LLMs choke on long numerical sequences, but a simple separator token trick can boost accuracy by 35% and cut token costs by 16%—without any training.
Today's best AI agents fail at realistic software engineering tasks, stalling before even reaching 30% completion, revealing the urgent need for better long-horizon planning and human-AI collaboration.
LLMs are surprisingly bad at keeping up with how people's minds change over time, lagging humans by 45% on a new benchmark designed to test this crucial social skill.