Update README.md
Browse files
README.md
CHANGED
@@ -6,6 +6,8 @@ tags:
|
|
6 |
model-index:
|
7 |
- name: Code-Llama-3-8B-finetuned-py-to-cpp
|
8 |
results: []
|
|
|
|
|
9 |
---
|
10 |
|
11 |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
@@ -61,23 +63,4 @@ The following hyperparameters were used during training:
|
|
61 |
- Datasets 3.0.1
|
62 |
- Tokenizers 0.13.3
|
63 |
|
64 |
-
|
65 |
-
|
66 |
-
```
|
67 |
-
from transformers import AutoModelForCausalLM
|
68 |
-
from transformers import BitsAndBytesConfig
|
69 |
-
import torch
|
70 |
-
|
71 |
-
quantization_config = BitsAndBytesConfig(
|
72 |
-
load_in_4bit=True,
|
73 |
-
bnb_4bit_quant_type="nf4",
|
74 |
-
bnb_4bit_use_double_quant=True,
|
75 |
-
bnb_4bit_compute_dtype=torch.bfloat16,
|
76 |
-
)
|
77 |
-
|
78 |
-
model = AutoModelForCausalLM.from_pretrained(
|
79 |
-
"hugo-albert/Code-Llama-3-8B-finetuned-py-to-cpp",
|
80 |
-
quantization_config=quantization_config,
|
81 |
-
torch_dtype=torch.bfloat16,
|
82 |
-
)
|
83 |
-
```
|
|
|
6 |
model-index:
|
7 |
- name: Code-Llama-3-8B-finetuned-py-to-cpp
|
8 |
results: []
|
9 |
+
pipeline_tag: text-generation
|
10 |
+
library_name: transformers
|
11 |
---
|
12 |
|
13 |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
|
|
63 |
- Datasets 3.0.1
|
64 |
- Tokenizers 0.13.3
|
65 |
|
66 |
+
```
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|