apepkuss79
commited on
Update README.md
Browse files
README.md
CHANGED
@@ -37,7 +37,7 @@ tags:
|
|
37 |
|
38 |
## Run with LlamaEdge
|
39 |
|
40 |
-
- LlamaEdge version: [v0.
|
41 |
|
42 |
- Prompt template
|
43 |
|
@@ -126,4 +126,4 @@ tags:
|
|
126 |
| [Llama-3.2-3B-Instruct-Q8_0.gguf](https://huggingface.co/second-state/Llama-3.2-3B-Instruct-GGUF/blob/main/Llama-3.2-3B-Instruct-Q8_0.gguf) | Q8_0 | 8 | 1.32 GB| very large, extremely low quality loss - not recommended |
|
127 |
| [Llama-3.2-3B-Instruct-f16.gguf](https://huggingface.co/second-state/Llama-3.2-3B-Instruct-GGUF/blob/main/Llama-3.2-3B-Instruct-f16.gguf) | f16 | 16 | 2.48 GB| |
|
128 |
|
129 |
-
*Quantized with llama.cpp
|
|
|
37 |
|
38 |
## Run with LlamaEdge
|
39 |
|
40 |
+
- LlamaEdge version: [v0.16.5](https://github.com/LlamaEdge/LlamaEdge/releases/tag/0.16.5) and above
|
41 |
|
42 |
- Prompt template
|
43 |
|
|
|
126 |
| [Llama-3.2-3B-Instruct-Q8_0.gguf](https://huggingface.co/second-state/Llama-3.2-3B-Instruct-GGUF/blob/main/Llama-3.2-3B-Instruct-Q8_0.gguf) | Q8_0 | 8 | 1.32 GB| very large, extremely low quality loss - not recommended |
|
127 |
| [Llama-3.2-3B-Instruct-f16.gguf](https://huggingface.co/second-state/Llama-3.2-3B-Instruct-GGUF/blob/main/Llama-3.2-3B-Instruct-f16.gguf) | f16 | 16 | 2.48 GB| |
|
128 |
|
129 |
+
*Quantized with llama.cpp b4466*
|