Mistral logo

Mistral

Mistral 7B Instruct v0.1

A 7.3B parameter model that outperforms Llama 2 13B on all benchmarks, with optimizations for speed and context length.

Input / 1M tokens
$0.110
Output / 1M tokens
$0.190
Context window
3K tokens
Provider
Mistral
Knowledge cutoff
2023-09-30

Performance

Median streaming throughput and first-token latency measured by Artificial Analysis.

Output tokens / sec
Time to first token