lbourdois commited on
Commit
72e08df
·
verified ·
1 Parent(s): 5c34ad1

Improve language tag

Browse files

Hi! As the model is multilingual, this is a PR to add other languages than English to the language tag to improve the referencing. Note that 29 languages are announced in the README, but only 13 are explicitly listed. I was therefore only able to add these 13 languages.

Files changed (1) hide show
  1. README.md +61 -47
README.md CHANGED
@@ -1,47 +1,61 @@
1
- ---
2
- base_model:
3
- - prithivMLmods/Deepthink-Reasoning-7B
4
- - prithivMLmods/Neumind-Math-7B-Instruct
5
- - Qwen/Qwen2.5-7B-Instruct
6
- - prithivMLmods/QwQ-LCoT-7B-Instruct
7
- - prithivMLmods/QwQ-MathOct-7B
8
- library_name: transformers
9
- tags:
10
- - mergekit
11
- - merge
12
- ---
13
- ![omni.png](https://cdn-uploads.huggingface.co/production/uploads/65bb837dbfb878f46c77de4c/VodPeXQO9_KS-wg_R8vKW.png)
14
-
15
- # **Merge Qwen Based**
16
-
17
- This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
18
-
19
- ## Merge Details
20
- ### Merge Method
21
-
22
- This model was merged using the [Model Stock](https://arxiv.org/abs/2403.19522) merge method using [Qwen/Qwen2.5-7B-Instruct](https://huggingface.co/Qwen/Qwen2.5-7B-Instruct) as a base.
23
-
24
- ### Models Merged
25
-
26
- The following models were included in the merge:
27
- * [prithivMLmods/Deepthink-Reasoning-7B](https://huggingface.co/prithivMLmods/Deepthink-Reasoning-7B)
28
- * [prithivMLmods/Neumind-Math-7B-Instruct](https://huggingface.co/prithivMLmods/Neumind-Math-7B-Instruct)
29
- * [prithivMLmods/QwQ-LCoT-7B-Instruct](https://huggingface.co/prithivMLmods/QwQ-LCoT-7B-Instruct)
30
- * [prithivMLmods/QwQ-MathOct-7B](https://huggingface.co/prithivMLmods/QwQ-MathOct-7B)
31
-
32
- ### Configuration
33
-
34
- The following YAML configuration was used to produce this model:
35
-
36
- ```yaml
37
- models:
38
- - model: prithivMLmods/Deepthink-Reasoning-7B
39
- - model: prithivMLmods/QwQ-LCoT-7B-Instruct
40
- - model: prithivMLmods/QwQ-MathOct-7B
41
- - model: prithivMLmods/Neumind-Math-7B-Instruct
42
- merge_method: model_stock
43
- base_model: Qwen/Qwen2.5-7B-Instruct
44
- normalize: true
45
- int8_mask: true
46
- dtype: bfloat16
47
- ```
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model:
3
+ - prithivMLmods/Deepthink-Reasoning-7B
4
+ - prithivMLmods/Neumind-Math-7B-Instruct
5
+ - Qwen/Qwen2.5-7B-Instruct
6
+ - prithivMLmods/QwQ-LCoT-7B-Instruct
7
+ - prithivMLmods/QwQ-MathOct-7B
8
+ library_name: transformers
9
+ tags:
10
+ - mergekit
11
+ - merge
12
+ language:
13
+ - zho
14
+ - eng
15
+ - fra
16
+ - spa
17
+ - por
18
+ - deu
19
+ - ita
20
+ - rus
21
+ - jpn
22
+ - kor
23
+ - vie
24
+ - tha
25
+ - ara
26
+ ---
27
+ ![omni.png](https://cdn-uploads.huggingface.co/production/uploads/65bb837dbfb878f46c77de4c/VodPeXQO9_KS-wg_R8vKW.png)
28
+
29
+ # **Merge Qwen Based**
30
+
31
+ This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
32
+
33
+ ## Merge Details
34
+ ### Merge Method
35
+
36
+ This model was merged using the [Model Stock](https://arxiv.org/abs/2403.19522) merge method using [Qwen/Qwen2.5-7B-Instruct](https://huggingface.co/Qwen/Qwen2.5-7B-Instruct) as a base.
37
+
38
+ ### Models Merged
39
+
40
+ The following models were included in the merge:
41
+ * [prithivMLmods/Deepthink-Reasoning-7B](https://huggingface.co/prithivMLmods/Deepthink-Reasoning-7B)
42
+ * [prithivMLmods/Neumind-Math-7B-Instruct](https://huggingface.co/prithivMLmods/Neumind-Math-7B-Instruct)
43
+ * [prithivMLmods/QwQ-LCoT-7B-Instruct](https://huggingface.co/prithivMLmods/QwQ-LCoT-7B-Instruct)
44
+ * [prithivMLmods/QwQ-MathOct-7B](https://huggingface.co/prithivMLmods/QwQ-MathOct-7B)
45
+
46
+ ### Configuration
47
+
48
+ The following YAML configuration was used to produce this model:
49
+
50
+ ```yaml
51
+ models:
52
+ - model: prithivMLmods/Deepthink-Reasoning-7B
53
+ - model: prithivMLmods/QwQ-LCoT-7B-Instruct
54
+ - model: prithivMLmods/QwQ-MathOct-7B
55
+ - model: prithivMLmods/Neumind-Math-7B-Instruct
56
+ merge_method: model_stock
57
+ base_model: Qwen/Qwen2.5-7B-Instruct
58
+ normalize: true
59
+ int8_mask: true
60
+ dtype: bfloat16
61
+ ```