fsaudm commited on
Commit
253471f
1 Parent(s): 91cebb2

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +40 -5
README.md CHANGED
@@ -1,21 +1,56 @@
1
  ---
2
  library_name: transformers
3
  base_model: meta-llama/Meta-Llama-3.1-8B-Instruct
4
- license: llama3
5
  model-index:
6
  - name: Meta-Llama-3.1-8B-Instruct-INT4
7
  results: []
8
  language:
9
  - en
10
- - es
 
11
  - it
12
- - ar
 
 
 
 
 
 
 
 
 
13
  ---
14
 
15
  # Model Card for Model ID
16
 
17
  <!-- Provide a quick summary of what the model is/does. -->
 
 
 
18
  - **Developed by:** [More Information Needed]
19
- - **License:** [More Information Needed]
20
- - **Finetuned from model [optional]:** [More Information Needed]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
21
 
 
1
  ---
2
  library_name: transformers
3
  base_model: meta-llama/Meta-Llama-3.1-8B-Instruct
4
+ license: llama3.1
5
  model-index:
6
  - name: Meta-Llama-3.1-8B-Instruct-INT4
7
  results: []
8
  language:
9
  - en
10
+ - de
11
+ - fr
12
  - it
13
+ - pt
14
+ - hi
15
+ - es
16
+ - th
17
+ tags:
18
+ - facebook
19
+ - meta
20
+ - pytorch
21
+ - llama
22
+ - llama-3
23
  ---
24
 
25
  # Model Card for Model ID
26
 
27
  <!-- Provide a quick summary of what the model is/does. -->
28
+
29
+ This is a quantized version of `Llama 3.1 70B Instruct`. Quantization to **4-bit** using `bistandbytes` and `accelerate`.
30
+
31
  - **Developed by:** [More Information Needed]
32
+ - **License:** llama3.1
33
+ - **Base Model [optional]:** meta-llama/Meta-Llama-3.1-8B-Instruct
34
+
35
+ ```
36
+ # Use a pipeline as a high-level helper
37
+ from transformers import pipeline
38
+
39
+ messages = [
40
+ {"role": "user", "content": "Who are you?"},
41
+ ]
42
+ pipe = pipeline("text-generation", model="meta-llama/Meta-Llama-3.1-8B-Instruct")
43
+ pipe(messages) Copy # Load model directly
44
+ ```
45
+
46
+ ```
47
+ from transformers import AutoTokenizer, AutoModelForCausalLM
48
+
49
+ tokenizer = AutoTokenizer.from_pretrained("meta-llama/Meta-Llama-3.1-8B-Instruct")
50
+ model = AutoModelForCausalLM.from_pretrained("meta-llama/Meta-Llama-3.1-8B-Instruct")
51
+ ```
52
+
53
+ The model information can be found in the original [meta-llama/Meta-Llama-3.1-8B-Instruct](https://huggingface.co/meta-llama/Meta-Llama-3.1-8B-Instruct)
54
+
55
+
56