Mistral
Mistral 7B Instruct v0.1
A 7.3B parameter model that outperforms Llama 2 13B on all benchmarks, with optimizations for speed and context length.
- Input / 1M tokens
- $0.110
- Output / 1M tokens
- $0.190
- Context window
- 3K tokens
- Provider
- Mistral
- Knowledge cutoff
- 2023-09-30
Performance
Median streaming throughput and first-token latency measured by Artificial Analysis.
- Output tokens / sec
- —
- Time to first token
- —