Crystalcareai
commited on
Update README.md
Browse files
README.md
CHANGED
@@ -28,6 +28,9 @@ This model utilizes PEFT layer replication at inference time to duplicate layers
|
|
28 |
and the adapter that is attached as well. Performance will be similar with both methods, but VRAM use is considerably less when using the adapter.
|
29 |
This model was initialized using [Unsloth's Mistralfied Phi-3-Instruct-4k](https://huggingface.co/unsloth/Phi-3-mini-4k-instruct). If you choose to use the adapter method, please attach it their model.
|
30 |
|
|
|
|
|
|
|
31 |
This model is based on Phi-3-Mini-Instruct-4k, and is governed by the MIT license in which Microsoft released Phi-3.
|
32 |
|
33 |
The base model has 4k context, and the qLoRA fine-tuning was with 4k sequence length.
|
|
|
28 |
and the adapter that is attached as well. Performance will be similar with both methods, but VRAM use is considerably less when using the adapter.
|
29 |
This model was initialized using [Unsloth's Mistralfied Phi-3-Instruct-4k](https://huggingface.co/unsloth/Phi-3-mini-4k-instruct). If you choose to use the adapter method, please attach it their model.
|
30 |
|
31 |
+
![Image](https://i.ibb.co/C6sqLBH/Vram-Use.png)
|
32 |
+
|
33 |
+
|
34 |
This model is based on Phi-3-Mini-Instruct-4k, and is governed by the MIT license in which Microsoft released Phi-3.
|
35 |
|
36 |
The base model has 4k context, and the qLoRA fine-tuning was with 4k sequence length.
|