Search papers, labs, and topics across Lattice.
This paper introduces gpt-oss-120b and gpt-oss-20b, two open-weight reasoning models built using a mixture-of-experts transformer architecture and trained via large-scale distillation and reinforcement learning. These models are optimized for agentic capabilities, including research browsing and tool use, and utilize a chat format for instruction following. The authors demonstrate strong performance on mathematics, coding, and safety benchmarks and release the model weights and related resources under an Apache 2.0 license.
Open-weight reasoning models now rival proprietary systems in agentic capabilities and benchmark performance, thanks to gpt-oss-120b and gpt-oss-20b.
We present gpt-oss-120b and gpt-oss-20b, two open-weight reasoning models that push the frontier of accuracy and inference cost. The models use an efficient mixture-of-expert transformer architecture and are trained using large-scale distillation and reinforcement learning. We optimize the models to have strong agentic capabilities (deep research browsing, python tool use, and support for developer-provided functions), all while using a rendered chat format that enables clear instruction following and role delineation. Both models achieve strong results on benchmarks ranging from mathematics, coding, and safety. We release the model weights, inference implementations, tool environments, and tokenizers under an Apache 2.0 license to enable broad use and further research.