Supahands commited on
Commit
231a547
·
verified ·
1 Parent(s): bac811b

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +48 -0
README.md ADDED
@@ -0,0 +1,48 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - ms
4
+ tags:
5
+ - llama-cpp
6
+ - gguf
7
+ - quantized
8
+ - llm
9
+ - llama
10
+ - ollama
11
+ base_model: mesolitica/llama-7b-hf-32768-fpf
12
+ ---
13
+
14
+ # Supa-AI/llama-7b-hf-32768-fpf-gguf
15
+ This model was converted to GGUF format from [`mesolitica/llama-7b-hf-32768-fpf`](https://huggingface.co/mesolitica/llama-7b-hf-32768-fpf) using llama.cpp.
16
+ Refer to the [original model card](https://huggingface.co/mesolitica/llama-7b-hf-32768-fpf) for more details on the model.
17
+
18
+ ## Available Versions
19
+ - `llama-7b-hf-32768-fpf.q4_0.gguf` (q4_0)
20
+ - `llama-7b-hf-32768-fpf.q4_1.gguf` (q4_1)
21
+ - `llama-7b-hf-32768-fpf.q5_0.gguf` (q5_0)
22
+ - `llama-7b-hf-32768-fpf.q5_1.gguf` (q5_1)
23
+ - `llama-7b-hf-32768-fpf.q8_0.gguf` (q8_0)
24
+ - `llama-7b-hf-32768-fpf.q3_k_s.gguf` (q3_K_S)
25
+ - `llama-7b-hf-32768-fpf.q3_k_m.gguf` (q3_K_M)
26
+ - `llama-7b-hf-32768-fpf.q3_k_l.gguf` (q3_K_L)
27
+ - `llama-7b-hf-32768-fpf.q4_k_s.gguf` (q4_K_S)
28
+ - `llama-7b-hf-32768-fpf.q4_k_m.gguf` (q4_K_M)
29
+ - `llama-7b-hf-32768-fpf.q5_k_s.gguf` (q5_K_S)
30
+ - `llama-7b-hf-32768-fpf.q5_k_m.gguf` (q5_K_M)
31
+ - `llama-7b-hf-32768-fpf.q6_k.gguf` (q6_K)
32
+
33
+ ## Use with llama.cpp
34
+ Replace `FILENAME` with one of the above filenames.
35
+
36
+ ### CLI:
37
+ ```bash
38
+ llama-cli --hf-repo Supa-AI/llama-7b-hf-32768-fpf-gguf --hf-file FILENAME -p "Your prompt here"
39
+ ```
40
+
41
+ ### Server:
42
+ ```bash
43
+ llama-server --hf-repo Supa-AI/llama-7b-hf-32768-fpf-gguf --hf-file FILENAME -c 2048
44
+ ```
45
+
46
+ ## Model Details
47
+ - **Original Model:** [mesolitica/llama-7b-hf-32768-fpf](https://huggingface.co/mesolitica/llama-7b-hf-32768-fpf)
48
+ - **Format:** GGUF