Add pipeline tag, paper link, project page link and Github repo

#18
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +11 -8
README.md CHANGED
@@ -1,4 +1,6 @@
1
  ---
 
 
2
  language:
3
  - en
4
  - fr
@@ -24,14 +26,12 @@ language:
24
  - vi
25
  - hi
26
  - bn
27
- license: apache-2.0
28
  library_name: vllm
 
29
  inference: false
30
- base_model:
31
- - mistralai/Mistral-Small-3.1-24B-Instruct-2503
32
- extra_gated_description: >-
33
- If you want to learn more about how we process your personal data, please read
34
- our <a href="https://mistral.ai/terms/">Privacy Policy</a>.
35
  ---
36
 
37
  # Model Card for Magistral-Small-2506
@@ -42,9 +42,11 @@ Magistral Small can be deployed locally, fitting within a single RTX 4090 or a 3
42
 
43
  Learn more about Magistral in our [blog post](https://mistral.ai/news/magistral/).
44
 
 
 
45
  ## Key Features
46
  - **Reasoning:** Capable of long chains of reasoning traces before providing an answer.
47
- - **Multilingual:** Supports dozens of languages, including English, French, German, Greek, Hindi, Indonesian, Italian, Japanese, Korean, Malay, Nepali, Polish, Portuguese, Romanian, Russian, Serbian, Spanish, Swedish, Turkish, Ukrainian, Vietnamese, Arabic, Bengali, Chinese, and Farsi.
48
  - **Apache 2.0 License:** Open license allowing usage and modification for both commercial and non-commercial purposes.
49
  - **Context Window:** A 128k context window, **but** performance might degrade past **40k**. Hence we recommend setting the maximum model length to 40k.
50
 
@@ -208,7 +210,8 @@ for chunk in stream:
208
  if content is not None:
209
  if not printed_content:
210
  printed_content = True
211
- print("\ncontent:", end="", flush=True)
 
212
  # Extract and print the content
213
  print(content, end="", flush=True)
214
 
 
1
  ---
2
+ base_model:
3
+ - mistralai/Mistral-Small-3.1-24B-Instruct-2503
4
  language:
5
  - en
6
  - fr
 
26
  - vi
27
  - hi
28
  - bn
 
29
  library_name: vllm
30
+ license: apache-2.0
31
  inference: false
32
+ extra_gated_description: If you want to learn more about how we process your personal
33
+ data, please read our <a href="https://mistral.ai/terms/">Privacy Policy</a>.
34
+ pipeline_tag: text-generation
 
 
35
  ---
36
 
37
  # Model Card for Magistral-Small-2506
 
42
 
43
  Learn more about Magistral in our [blog post](https://mistral.ai/news/magistral/).
44
 
45
+ The model was presented in the paper [Magistral](https://huggingface.co/papers/2506.10910).
46
+
47
  ## Key Features
48
  - **Reasoning:** Capable of long chains of reasoning traces before providing an answer.
49
+ - **Multilingual:** Supports dozens of languages, including English, French, German, Greek, Hindi, Indonesian, Italian, Japanese, Korean, Malay, Nepali, Polish, Portuguese, Romanian, Russian, Serbian, Spanish, Turkish, Ukrainian, Vietnamese, Arabic, Bengali, Chinese, and Farsi.
50
  - **Apache 2.0 License:** Open license allowing usage and modification for both commercial and non-commercial purposes.
51
  - **Context Window:** A 128k context window, **but** performance might degrade past **40k**. Hence we recommend setting the maximum model length to 40k.
52
 
 
210
  if content is not None:
211
  if not printed_content:
212
  printed_content = True
213
+ print("
214
+ content:", end="", flush=True)
215
  # Extract and print the content
216
  print(content, end="", flush=True)
217