root
initial
a841f20
metadata
license: mit
datasets:
  - Vikhrmodels/GrandMaster-PRO-MAX
language:
  - en
  - ru
tags:
  - mistral
  - chat
  - conversational
  - transformers
inference:
  parameters:
    temperature: 0
pipeline_tag: text-generation
base_model:
  - mistralai/Mistral-Small-3.1-24B-Instruct-2503
  - anthracite-core/Mistral-Small-3.1-24B-Instruct-2503-HF
library_name: vllm

Zero-Mistral-Small-3.1-24B-Instruct-2503-beta

Zero-Mistral-Small-3.1 is an improved TEXT-ONLY version of mistralai/Mistral-Small-3.1-24B-Instruct-2503 based on NO-VISION anthracite-core/Mistral-Small-3.1-24B-Instruct-2503-HF, primarily adapted for Russian and English languages. The training involved SFT stage on GrandMaster-PRO-MAX dataset.

This is a beta version. Benchmarks and some more fine-tuning coming soon.

Current status:

Trained with lm_head
Train loss: 0.564200
Eval loss: 0.638504