Zac Zuo

OLMo 2 32B - First fully open SOTA model from Ai2

byโ€ข

OLMo 2 32B is the most capable and largest model in the OLMo 2 family. High-performance, fully open LLM. Outperforms GPT-3.5 Turbo & GPT-4o mini. Trained efficiently on 6T tokens.

Add a comment

Replies

Best
Zac Zuo
Hunter
๐Ÿ“Œ

Hi everyone!

Sharing OLMo 2 32B, a new language model from the Allen Institute for AI (Ai2). This model is notable for being fully open, with the data, code, and weights all publicly available.

Key features:

๐ŸŽ›๏ธ 32B parameters, placing it in a mid-size range for current LLMs.
๐Ÿง  Trained on 6 trillion tokens, using a two-stage pretraining process.
๐Ÿ“Š AI2 reports performance comparable to GPT-3.5 Turbo and GPT-4o mini on several benchmarks.
โš™๏ธ The training process incorporated techniques like Group Relative Policy Optimization (GRPO) and RLVR.
๐Ÿ”“ The model, training data, and code are available under an Apache 2.0 license.

This release provides researchers and developers with full access to a capable LLM and its training pipeline. You have to admire Ai2's commitment to transparency here.

Try it by yourself in Ai2's Playground.