lbourdois commited on
Commit
28dec9e
·
verified ·
1 Parent(s): 7c4ee5d

Improve language tag

Browse files

Hi! As the model is multilingual, this is a PR to add other languages than English to the language tag to improve the referencing. Note that 29 languages are announced in the README, but only 13 are explicitly listed. I was therefore only able to add these 13 languages.

Files changed (1) hide show
  1. README.md +130 -118
README.md CHANGED
@@ -1,119 +1,131 @@
1
- ---
2
- license: apache-2.0
3
- language:
4
- - en
5
- base_model:
6
- - Qwen/Qwen2.5-72B-Instruct
7
- pipeline_tag: text-generation
8
- library_name: transformers
9
- tags:
10
- - reasoning
11
- - logic
12
- - cot
13
- - text-generation-inference
14
- new_version: Daemontatox/Cogito-Maximus
15
- ---
16
-
17
- ![image](./image.webp)
18
-
19
-
20
- ## **Model Overview**
21
-
22
- This model, **Cogito-Maximus**, is a fine-tuned version of the `unsloth/qwen2.5-72b-instruct-bnb-4bit` base model, optimized for advanced text generation tasks. It leverages the power of **Unsloth** and **Huggingface's TRL (Transformer Reinforcement Learning)** library to achieve faster training and improved performance.
23
-
24
- ### **Key Features**
25
- - **Base Model:** `unsloth/qwen2.5-72b-instruct`
26
- - **Training Acceleration:** Trained 2x faster using [Unsloth](https://github.com/unslothai/unsloth).
27
- - **Fine-Tuning Framework:** Utilizes Huggingface's [TRL](https://github.com/huggingface/trl) library.
28
- - **Optimized for Inference:** Ready for deployment in text-generation tasks with efficient inference capabilities.
29
- - **License:** Apache-2.0
30
-
31
- ---
32
-
33
- ## **Model Details**
34
-
35
- ### **Developed by**
36
- - **Author:** Daemontatox
37
- - **Organization:** Independent Contributor
38
-
39
- ### **Tags**
40
- - Text Generation Inference
41
- - Transformers
42
- - Unsloth
43
- - Qwen2
44
- - TRL
45
-
46
- ### **Language**
47
- - English (`en`)
48
-
49
- ### **License**
50
- This model is released under the **Apache-2.0 License**, which allows for free use, modification, and distribution, provided the original license and copyright notice are included.
51
-
52
- ---
53
-
54
- ## **Model Training**
55
-
56
- ### **Base Model**
57
- The model is derived from the `unsloth/qwen2.5-72b-instruct`, a version of the Qwen2.5-72B instruction-tuned model. The base model is optimized for efficiency using **bitsandbytes (bnb)** 4-bit quantization.
58
-
59
- ### **Training Process**
60
- - **Framework:** The model was fine-tuned using **Unsloth**, a library designed to accelerate the training of large language models.
61
- - **Acceleration:** Training was completed **2x faster** compared to traditional methods, thanks to Unsloth's optimizations.
62
- - **Reinforcement Learning:** Fine-tuning incorporated techniques from Huggingface's **TRL** library, enabling advanced instruction-tuning and alignment with human preferences.
63
-
64
- ---
65
-
66
- ## **Intended Use**
67
-
68
- ### **Primary Use Case**
69
- This model is designed for **text generation tasks**, including but not limited to:
70
- - Instruction-following
71
- - Question answering
72
- - Content creation
73
- - Dialogue systems
74
-
75
- ### **Limitations**
76
- - The model is trained primarily on English data and may not perform as well on other languages.
77
- - While fine-tuned for instruction-following, outputs should be reviewed for accuracy and relevance in critical applications.
78
-
79
- ---
80
-
81
- ## **How to Use**
82
-
83
- ### **Installation**
84
- To use this model, ensure you have the following libraries installed:
85
- ```bash
86
- pip install transformers torch bitsandbytes unsloth trl
87
- ```
88
-
89
-
90
-
91
-
92
- ```python
93
- from transformers import AutoModelForCausalLM, AutoTokenizer
94
-
95
- # Load the tokenizer and model
96
- model_name = "Daemontatox/Cogito-Maximus"
97
- tokenizer = AutoTokenizer.from_pretrained(model_name)
98
- model = AutoModelForCausalLM.from_pretrained(model_name, device_map="auto", load_in_4bit=True)
99
-
100
- # Generate text
101
- input_text = "Explain the concept of machine learning in simple terms."
102
- inputs = tokenizer(input_text, return_tensors="pt").to("cuda")
103
- outputs = model.generate(**inputs, max_length=100)
104
-
105
- # Decode and print the output
106
- print(tokenizer.decode(outputs[0], skip_special_tokens=True))
107
-
108
- ```
109
-
110
- ```
111
- @misc{daemontatox_cogito_maximus,
112
- author = {Daemontatox},
113
- title = {Cogito-Maximus: Fine-tuned Qwen2.5-72B Instruct Model},
114
- year = {2025},
115
- publisher = {Hugging Face},
116
- journal = {Hugging Face Model Repository},
117
- howpublished = {\url{https://huggingface.co/Daemontatox/Cogito-Maximus}}
118
- }
 
 
 
 
 
 
 
 
 
 
 
 
119
  ```
 
1
+ ---
2
+ license: apache-2.0
3
+ language:
4
+ - zho
5
+ - eng
6
+ - fra
7
+ - spa
8
+ - por
9
+ - deu
10
+ - ita
11
+ - rus
12
+ - jpn
13
+ - kor
14
+ - vie
15
+ - tha
16
+ - ara
17
+ base_model:
18
+ - Qwen/Qwen2.5-72B-Instruct
19
+ pipeline_tag: text-generation
20
+ library_name: transformers
21
+ tags:
22
+ - reasoning
23
+ - logic
24
+ - cot
25
+ - text-generation-inference
26
+ new_version: Daemontatox/Cogito-Maximus
27
+ ---
28
+
29
+ ![image](./image.webp)
30
+
31
+
32
+ ## **Model Overview**
33
+
34
+ This model, **Cogito-Maximus**, is a fine-tuned version of the `unsloth/qwen2.5-72b-instruct-bnb-4bit` base model, optimized for advanced text generation tasks. It leverages the power of **Unsloth** and **Huggingface's TRL (Transformer Reinforcement Learning)** library to achieve faster training and improved performance.
35
+
36
+ ### **Key Features**
37
+ - **Base Model:** `unsloth/qwen2.5-72b-instruct`
38
+ - **Training Acceleration:** Trained 2x faster using [Unsloth](https://github.com/unslothai/unsloth).
39
+ - **Fine-Tuning Framework:** Utilizes Huggingface's [TRL](https://github.com/huggingface/trl) library.
40
+ - **Optimized for Inference:** Ready for deployment in text-generation tasks with efficient inference capabilities.
41
+ - **License:** Apache-2.0
42
+
43
+ ---
44
+
45
+ ## **Model Details**
46
+
47
+ ### **Developed by**
48
+ - **Author:** Daemontatox
49
+ - **Organization:** Independent Contributor
50
+
51
+ ### **Tags**
52
+ - Text Generation Inference
53
+ - Transformers
54
+ - Unsloth
55
+ - Qwen2
56
+ - TRL
57
+
58
+ ### **Language**
59
+ - English (`en`)
60
+
61
+ ### **License**
62
+ This model is released under the **Apache-2.0 License**, which allows for free use, modification, and distribution, provided the original license and copyright notice are included.
63
+
64
+ ---
65
+
66
+ ## **Model Training**
67
+
68
+ ### **Base Model**
69
+ The model is derived from the `unsloth/qwen2.5-72b-instruct`, a version of the Qwen2.5-72B instruction-tuned model. The base model is optimized for efficiency using **bitsandbytes (bnb)** 4-bit quantization.
70
+
71
+ ### **Training Process**
72
+ - **Framework:** The model was fine-tuned using **Unsloth**, a library designed to accelerate the training of large language models.
73
+ - **Acceleration:** Training was completed **2x faster** compared to traditional methods, thanks to Unsloth's optimizations.
74
+ - **Reinforcement Learning:** Fine-tuning incorporated techniques from Huggingface's **TRL** library, enabling advanced instruction-tuning and alignment with human preferences.
75
+
76
+ ---
77
+
78
+ ## **Intended Use**
79
+
80
+ ### **Primary Use Case**
81
+ This model is designed for **text generation tasks**, including but not limited to:
82
+ - Instruction-following
83
+ - Question answering
84
+ - Content creation
85
+ - Dialogue systems
86
+
87
+ ### **Limitations**
88
+ - The model is trained primarily on English data and may not perform as well on other languages.
89
+ - While fine-tuned for instruction-following, outputs should be reviewed for accuracy and relevance in critical applications.
90
+
91
+ ---
92
+
93
+ ## **How to Use**
94
+
95
+ ### **Installation**
96
+ To use this model, ensure you have the following libraries installed:
97
+ ```bash
98
+ pip install transformers torch bitsandbytes unsloth trl
99
+ ```
100
+
101
+
102
+
103
+
104
+ ```python
105
+ from transformers import AutoModelForCausalLM, AutoTokenizer
106
+
107
+ # Load the tokenizer and model
108
+ model_name = "Daemontatox/Cogito-Maximus"
109
+ tokenizer = AutoTokenizer.from_pretrained(model_name)
110
+ model = AutoModelForCausalLM.from_pretrained(model_name, device_map="auto", load_in_4bit=True)
111
+
112
+ # Generate text
113
+ input_text = "Explain the concept of machine learning in simple terms."
114
+ inputs = tokenizer(input_text, return_tensors="pt").to("cuda")
115
+ outputs = model.generate(**inputs, max_length=100)
116
+
117
+ # Decode and print the output
118
+ print(tokenizer.decode(outputs[0], skip_special_tokens=True))
119
+
120
+ ```
121
+
122
+ ```
123
+ @misc{daemontatox_cogito_maximus,
124
+ author = {Daemontatox},
125
+ title = {Cogito-Maximus: Fine-tuned Qwen2.5-72B Instruct Model},
126
+ year = {2025},
127
+ publisher = {Hugging Face},
128
+ journal = {Hugging Face Model Repository},
129
+ howpublished = {\url{https://huggingface.co/Daemontatox/Cogito-Maximus}}
130
+ }
131
  ```