Update README.md
Browse files
README.md
CHANGED
@@ -2,7 +2,7 @@
|
|
2 |
|
3 |
# Hicoder-R1-Distill-Gemma-27B
|
4 |
|
5 |
-
|
6 |
### Model Overview
|
7 |
|
8 |
**Hicoder-R1-Distill-Gemma-27B** is a large language model fine-tuned from Google's **Gemma-2 27B** (*Note: Assuming Gemma-2 27B as Gemma-3 is not publicly released*) base model. This model is specifically optimized for **Chain-of-Thought (CoT) reasoning** and **code generation** tasks. The "Distill" in the name suggests that knowledge distillation techniques may have been employed during the fine-tuning process, potentially leveraging outputs from a more powerful teacher model to enhance its reasoning and coding abilities.
|
|
|
2 |
|
3 |
# Hicoder-R1-Distill-Gemma-27B
|
4 |
|
5 |
+
Notably, this CoT-enabled model was trained using only a single RTX 4090D, achieved through optimizations in both GPU VRAM and system RAM management, as well as specific techniques applied during the training steps.
|
6 |
### Model Overview
|
7 |
|
8 |
**Hicoder-R1-Distill-Gemma-27B** is a large language model fine-tuned from Google's **Gemma-2 27B** (*Note: Assuming Gemma-2 27B as Gemma-3 is not publicly released*) base model. This model is specifically optimized for **Chain-of-Thought (CoT) reasoning** and **code generation** tasks. The "Distill" in the name suggests that knowledge distillation techniques may have been employed during the fine-tuning process, potentially leveraging outputs from a more powerful teacher model to enhance its reasoning and coding abilities.
|