tonyli8623 commited on
Commit
a41b938
·
verified ·
1 Parent(s): 76780a9

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -2,7 +2,7 @@
2
 
3
  # Hicoder-R1-Distill-Gemma-27B
4
 
5
-
6
  ### Model Overview
7
 
8
  **Hicoder-R1-Distill-Gemma-27B** is a large language model fine-tuned from Google's **Gemma-2 27B** (*Note: Assuming Gemma-2 27B as Gemma-3 is not publicly released*) base model. This model is specifically optimized for **Chain-of-Thought (CoT) reasoning** and **code generation** tasks. The "Distill" in the name suggests that knowledge distillation techniques may have been employed during the fine-tuning process, potentially leveraging outputs from a more powerful teacher model to enhance its reasoning and coding abilities.
 
2
 
3
  # Hicoder-R1-Distill-Gemma-27B
4
 
5
+ Notably, this CoT-enabled model was trained using only a single RTX 4090D, achieved through optimizations in both GPU VRAM and system RAM management, as well as specific techniques applied during the training steps.
6
  ### Model Overview
7
 
8
  **Hicoder-R1-Distill-Gemma-27B** is a large language model fine-tuned from Google's **Gemma-2 27B** (*Note: Assuming Gemma-2 27B as Gemma-3 is not publicly released*) base model. This model is specifically optimized for **Chain-of-Thought (CoT) reasoning** and **code generation** tasks. The "Distill" in the name suggests that knowledge distillation techniques may have been employed during the fine-tuning process, potentially leveraging outputs from a more powerful teacher model to enhance its reasoning and coding abilities.