Mistral
Mistral Nemo
A 12B parameter model with a 128k token context length built by Mistral in collaboration with NVIDIA. The model is multilingual, supporting English, French, German, Spanish, Italian, Portuguese, Chinese, Japanese, Korean, Arabic, and Hindi. It supports function calling and is released under the Apache 2.0 license.
- Input / 1M tokens
- $0.020
- Output / 1M tokens
- $0.040
- Context window
- 131K tokens
- Provider
- Mistral
- Knowledge cutoff
- 2024-04-30
Performance
Median streaming throughput and first-token latency measured by Artificial Analysis.
- Output tokens / sec
- —
- Time to first token
- —