Add pipeline tag, paper link, project page link and Github repo
#18
by
nielsr
HF Staff
- opened
README.md
CHANGED
@@ -1,4 +1,6 @@
|
|
1 |
---
|
|
|
|
|
2 |
language:
|
3 |
- en
|
4 |
- fr
|
@@ -24,14 +26,12 @@ language:
|
|
24 |
- vi
|
25 |
- hi
|
26 |
- bn
|
27 |
-
license: apache-2.0
|
28 |
library_name: vllm
|
|
|
29 |
inference: false
|
30 |
-
|
31 |
-
|
32 |
-
|
33 |
-
If you want to learn more about how we process your personal data, please read
|
34 |
-
our <a href="https://mistral.ai/terms/">Privacy Policy</a>.
|
35 |
---
|
36 |
|
37 |
# Model Card for Magistral-Small-2506
|
@@ -42,9 +42,11 @@ Magistral Small can be deployed locally, fitting within a single RTX 4090 or a 3
|
|
42 |
|
43 |
Learn more about Magistral in our [blog post](https://mistral.ai/news/magistral/).
|
44 |
|
|
|
|
|
45 |
## Key Features
|
46 |
- **Reasoning:** Capable of long chains of reasoning traces before providing an answer.
|
47 |
-
- **Multilingual:** Supports dozens of languages, including English, French, German, Greek, Hindi, Indonesian, Italian, Japanese, Korean, Malay, Nepali, Polish, Portuguese, Romanian, Russian, Serbian, Spanish,
|
48 |
- **Apache 2.0 License:** Open license allowing usage and modification for both commercial and non-commercial purposes.
|
49 |
- **Context Window:** A 128k context window, **but** performance might degrade past **40k**. Hence we recommend setting the maximum model length to 40k.
|
50 |
|
@@ -208,7 +210,8 @@ for chunk in stream:
|
|
208 |
if content is not None:
|
209 |
if not printed_content:
|
210 |
printed_content = True
|
211 |
-
print("
|
|
|
212 |
# Extract and print the content
|
213 |
print(content, end="", flush=True)
|
214 |
|
|
|
1 |
---
|
2 |
+
base_model:
|
3 |
+
- mistralai/Mistral-Small-3.1-24B-Instruct-2503
|
4 |
language:
|
5 |
- en
|
6 |
- fr
|
|
|
26 |
- vi
|
27 |
- hi
|
28 |
- bn
|
|
|
29 |
library_name: vllm
|
30 |
+
license: apache-2.0
|
31 |
inference: false
|
32 |
+
extra_gated_description: If you want to learn more about how we process your personal
|
33 |
+
data, please read our <a href="https://mistral.ai/terms/">Privacy Policy</a>.
|
34 |
+
pipeline_tag: text-generation
|
|
|
|
|
35 |
---
|
36 |
|
37 |
# Model Card for Magistral-Small-2506
|
|
|
42 |
|
43 |
Learn more about Magistral in our [blog post](https://mistral.ai/news/magistral/).
|
44 |
|
45 |
+
The model was presented in the paper [Magistral](https://huggingface.co/papers/2506.10910).
|
46 |
+
|
47 |
## Key Features
|
48 |
- **Reasoning:** Capable of long chains of reasoning traces before providing an answer.
|
49 |
+
- **Multilingual:** Supports dozens of languages, including English, French, German, Greek, Hindi, Indonesian, Italian, Japanese, Korean, Malay, Nepali, Polish, Portuguese, Romanian, Russian, Serbian, Spanish, Turkish, Ukrainian, Vietnamese, Arabic, Bengali, Chinese, and Farsi.
|
50 |
- **Apache 2.0 License:** Open license allowing usage and modification for both commercial and non-commercial purposes.
|
51 |
- **Context Window:** A 128k context window, **but** performance might degrade past **40k**. Hence we recommend setting the maximum model length to 40k.
|
52 |
|
|
|
210 |
if content is not None:
|
211 |
if not printed_content:
|
212 |
printed_content = True
|
213 |
+
print("
|
214 |
+
content:", end="", flush=True)
|
215 |
# Extract and print the content
|
216 |
print(content, end="", flush=True)
|
217 |
|