A
Allenai
Olmo 3.1 32B Instruct
Olmo 3.1 32B Instruct is a large-scale, 32-billion-parameter instruction-tuned language model engineered for high-performance conversational AI, multi-turn dialogue, and practical instruction following. As part of the Olmo 3.1 family, this variant emphasizes responsiveness to complex user directions and robust chat interactions while retaining strong capabilities on reasoning and coding benchmarks. Developed by Ai2 under the Apache 2.0 license, Olmo 3.1 32B Instruct reflects the Olmo initiative’s commitment to openness and transparency.
- Input / 1M tokens
- $0.200
- Output / 1M tokens
- $0.600
- Context window
- 66K tokens
- Provider
- Allenai
Performance
Median streaming throughput and first-token latency measured by Artificial Analysis.
- Output tokens / sec
- 54 t/s
- Time to first token
- 0.29s
Benchmarks
Intelligence, coding, and math indexes plus the underlying evaluation scores.
- Intelligence Index
- 12
- Coding Index
- 6
- Math Index
- —
- MMLU-Pro
- —
- GPQA
- 53.9%
- HLE
- 4.9%
- LiveCodeBench
- —
- SciCode
- 16.7%
- MATH-500
- —
- AIME
- —
Benchmarks via Artificial Analysis