MultivexAI commited on
Commit
0d2b9e6
·
verified ·
1 Parent(s): c34ac94

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -2
README.md CHANGED
@@ -6,7 +6,7 @@ base_model:
6
  - meta-llama/Llama-3.2-3B-Instruct
7
  ---
8
 
9
- # Gladiator-Mini-exp-1211: A Compact and Powerful Reasoning Engine
10
 
11
  **Gladiator-Mini-exp-1211** is a 3-billion parameter language model designed for **complex reasoning tasks**. This experimental model, based on [meta-llama/Llama-3.2-3B-Instruct] [https://huggingface.co/meta-llama/Llama-3.2-3B-Instruct], offers surprisingly strong analytical capabilities for its size. It demonstrates the potential of smaller models to achieve impressive performance in analytical thinking. We chose to finetune on a Llama model due to finetuning difficulties with the Qwen 2.5 3B model.
12
 
@@ -27,6 +27,8 @@ Gladiator-Mini-exp-1211 works best with **carefully crafted system prompts** tha
27
  **Why is it "Experimental"?**
28
 
29
  Gladiator-Mini-exp-1211 is still under development (hence the "exp" in its name). It may have biases, limitations, and areas where its reasoning is not yet perfect. We're releasing it to the community to gather feedback, explore its full capabilities, and continue to improve it. The experimental date is 12/11/2024.
 
 
30
 
31
  **The Goal:**
32
 
@@ -48,4 +50,6 @@ Gladiator-Mini-exp-1211 is an experimental model and should be used with caution
48
 
49
  Base model: https://huggingface.co/meta-llama/Llama-3.2-3B-Instruct
50
 
51
- Thanks to Meta for the fantastic Llama-3.2-3B model!
 
 
 
6
  - meta-llama/Llama-3.2-3B-Instruct
7
  ---
8
 
9
+ # Gladiator-Mini-exp-1211
10
 
11
  **Gladiator-Mini-exp-1211** is a 3-billion parameter language model designed for **complex reasoning tasks**. This experimental model, based on [meta-llama/Llama-3.2-3B-Instruct] [https://huggingface.co/meta-llama/Llama-3.2-3B-Instruct], offers surprisingly strong analytical capabilities for its size. It demonstrates the potential of smaller models to achieve impressive performance in analytical thinking. We chose to finetune on a Llama model due to finetuning difficulties with the Qwen 2.5 3B model.
12
 
 
27
  **Why is it "Experimental"?**
28
 
29
  Gladiator-Mini-exp-1211 is still under development (hence the "exp" in its name). It may have biases, limitations, and areas where its reasoning is not yet perfect. We're releasing it to the community to gather feedback, explore its full capabilities, and continue to improve it. The experimental date is 12/11/2024.
30
+ **Note: This is an experimental model, and its performance may vary compared to the base model. We are unsure.**
31
+
32
 
33
  **The Goal:**
34
 
 
50
 
51
  Base model: https://huggingface.co/meta-llama/Llama-3.2-3B-Instruct
52
 
53
+ Thanks to Meta for the fantastic Llama-3.2-3B model!
54
+
55
+ Newer experimental models will bring hopefully smarter models.