Edit model card

Model Card for Mistral-7B-v0.2

The Mistral-7B-Instruct-v0.2 Large Language Model (LLM) was fine-tuned on top of Mistral-7B-v0.2.

Mistral-7B-v0.2 has the following changes compared to Mistral-7B-v0.1

  • 32k context window (vs 8k context in v0.1)
  • Rope-theta = 1e6
  • No Sliding-Window Attention

For full details of this model please read our paper and release blog post.

Downloads last month
4
Safetensors
Model size
7.24B params
Tensor type
BF16
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for MaziyarPanahi/Mistral-7B-v0.2

Quantizations
1 model