lbourdois commited on
Commit
9946b05
·
verified ·
1 Parent(s): 81a36a2

Improve language tag

Browse files

Hi! As the model is multilingual, this is a PR to add other languages than English to the language tag to improve the referencing. Note that 29 languages are announced in the README, but only 13 are explicitly listed. I was therefore only able to add these 13 languages.

Files changed (1) hide show
  1. README.md +101 -90
README.md CHANGED
@@ -1,91 +1,102 @@
1
- ---
2
- license: other
3
- license_name: qwen-research
4
- license_link: https://huggingface.co/Qwen/Qwen2.5-3B/blob/main/LICENSE
5
- language:
6
- - fr
7
- - en
8
- pipeline_tag: text-generation
9
- tags:
10
- - chat
11
- - qwen
12
- - qwen2.5
13
- - finetune
14
- - french
15
- - legal
16
- - loi
17
- library_name: transformers
18
- inference: false
19
- model_creator: MaziyarPanahi
20
- quantized_by: MaziyarPanahi
21
- base_model: Qwen/Qwen2.5-3B
22
- model_name: calme-3.3-baguette-3b
23
- datasets:
24
- - MaziyarPanahi/calme-legalkit-v0.2
25
- ---
26
-
27
- <img src="./calme_3.png" alt="Calme-3 Models" width="800" style="margin-left:'auto' margin-right:'auto' display:'block'"/>
28
-
29
- > [!TIP]
30
- > This is avery small model, so it might not perform well for some prompts and may be sensitive to hyper parameters. I would appreciate any feedback to see if I can fix any issues in the next iteration. ❤️
31
-
32
-
33
- # MaziyarPanahi/calme-3.3-baguette-3b
34
-
35
- This model is an advanced iteration of the powerful `Qwen/Qwen2.5-3B`, specifically fine-tuned to enhance its capabilities in French Legal domain.
36
-
37
-
38
- # Quantized GGUF
39
-
40
- All GGUF models are available here: [MaziyarPanahi/calme-3.3-baguette-3b-GGUF](https://huggingface.co/MaziyarPanahi/calme-3.3-baguette-3b-GGUF)
41
-
42
-
43
- # 🏆 [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
44
-
45
- Leaderboard 2 coming soon!
46
-
47
-
48
- # Prompt Template
49
-
50
- This model uses `ChatML` prompt template:
51
-
52
- ```
53
- <|im_start|>system
54
- {System}
55
- <|im_end|>
56
- <|im_start|>user
57
- {User}
58
- <|im_end|>
59
- <|im_start|>assistant
60
- {Assistant}
61
- ````
62
-
63
- # How to use
64
-
65
-
66
- ```python
67
-
68
- # Use a pipeline as a high-level helper
69
-
70
- from transformers import pipeline
71
-
72
- messages = [
73
- {"role": "user", "content": "Who are you?"},
74
- ]
75
- pipe = pipeline("text-generation", model="MaziyarPanahi/calme-3.3-baguette-3b")
76
- pipe(messages)
77
-
78
-
79
- # Load model directly
80
-
81
- from transformers import AutoTokenizer, AutoModelForCausalLM
82
-
83
- tokenizer = AutoTokenizer.from_pretrained("MaziyarPanahi/calme-3.3-baguette-3b")
84
- model = AutoModelForCausalLM.from_pretrained("MaziyarPanahi/calme-3.3-baguette-3b")
85
- ```
86
-
87
-
88
-
89
- # Ethical Considerations
90
-
 
 
 
 
 
 
 
 
 
 
 
91
  As with any large language model, users should be aware of potential biases and limitations. We recommend implementing appropriate safeguards and human oversight when deploying this model in production environments.
 
1
+ ---
2
+ license: other
3
+ license_name: qwen-research
4
+ license_link: https://huggingface.co/Qwen/Qwen2.5-3B/blob/main/LICENSE
5
+ language:
6
+ - zho
7
+ - eng
8
+ - fra
9
+ - spa
10
+ - por
11
+ - deu
12
+ - ita
13
+ - rus
14
+ - jpn
15
+ - kor
16
+ - vie
17
+ - tha
18
+ - ara
19
+ pipeline_tag: text-generation
20
+ tags:
21
+ - chat
22
+ - qwen
23
+ - qwen2.5
24
+ - finetune
25
+ - french
26
+ - legal
27
+ - loi
28
+ library_name: transformers
29
+ inference: false
30
+ model_creator: MaziyarPanahi
31
+ quantized_by: MaziyarPanahi
32
+ base_model: Qwen/Qwen2.5-3B
33
+ model_name: calme-3.3-baguette-3b
34
+ datasets:
35
+ - MaziyarPanahi/calme-legalkit-v0.2
36
+ ---
37
+
38
+ <img src="./calme_3.png" alt="Calme-3 Models" width="800" style="margin-left:'auto' margin-right:'auto' display:'block'"/>
39
+
40
+ > [!TIP]
41
+ > This is avery small model, so it might not perform well for some prompts and may be sensitive to hyper parameters. I would appreciate any feedback to see if I can fix any issues in the next iteration. ❤️
42
+
43
+
44
+ # MaziyarPanahi/calme-3.3-baguette-3b
45
+
46
+ This model is an advanced iteration of the powerful `Qwen/Qwen2.5-3B`, specifically fine-tuned to enhance its capabilities in French Legal domain.
47
+
48
+
49
+ # ⚡ Quantized GGUF
50
+
51
+ All GGUF models are available here: [MaziyarPanahi/calme-3.3-baguette-3b-GGUF](https://huggingface.co/MaziyarPanahi/calme-3.3-baguette-3b-GGUF)
52
+
53
+
54
+ # 🏆 [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
55
+
56
+ Leaderboard 2 coming soon!
57
+
58
+
59
+ # Prompt Template
60
+
61
+ This model uses `ChatML` prompt template:
62
+
63
+ ```
64
+ <|im_start|>system
65
+ {System}
66
+ <|im_end|>
67
+ <|im_start|>user
68
+ {User}
69
+ <|im_end|>
70
+ <|im_start|>assistant
71
+ {Assistant}
72
+ ````
73
+
74
+ # How to use
75
+
76
+
77
+ ```python
78
+
79
+ # Use a pipeline as a high-level helper
80
+
81
+ from transformers import pipeline
82
+
83
+ messages = [
84
+ {"role": "user", "content": "Who are you?"},
85
+ ]
86
+ pipe = pipeline("text-generation", model="MaziyarPanahi/calme-3.3-baguette-3b")
87
+ pipe(messages)
88
+
89
+
90
+ # Load model directly
91
+
92
+ from transformers import AutoTokenizer, AutoModelForCausalLM
93
+
94
+ tokenizer = AutoTokenizer.from_pretrained("MaziyarPanahi/calme-3.3-baguette-3b")
95
+ model = AutoModelForCausalLM.from_pretrained("MaziyarPanahi/calme-3.3-baguette-3b")
96
+ ```
97
+
98
+
99
+
100
+ # Ethical Considerations
101
+
102
  As with any large language model, users should be aware of potential biases and limitations. We recommend implementing appropriate safeguards and human oversight when deploying this model in production environments.