Update README.md
Browse files
README.md
CHANGED
@@ -17,6 +17,8 @@ base_model: google/gemma-3-1b-it
|
|
17 |
**Original model**: [gemma-3-1b-it](https://huggingface.co/google/gemma-3-1b-it)<br>
|
18 |
**GGUF quantization:** provided by [bartowski](https://huggingface.co/bartowski) based on `llama.cpp` release [b4877](https://github.com/ggerganov/llama.cpp/releases/tag/b4877)<br>
|
19 |
|
|
|
|
|
20 |
## Technical Details
|
21 |
|
22 |
Supports a context length of 32k tokens, with a max output of 8192.
|
|
|
17 |
**Original model**: [gemma-3-1b-it](https://huggingface.co/google/gemma-3-1b-it)<br>
|
18 |
**GGUF quantization:** provided by [bartowski](https://huggingface.co/bartowski) based on `llama.cpp` release [b4877](https://github.com/ggerganov/llama.cpp/releases/tag/b4877)<br>
|
19 |
|
20 |
+
Requires llama.cpp runtime v1.19.0
|
21 |
+
|
22 |
## Technical Details
|
23 |
|
24 |
Supports a context length of 32k tokens, with a max output of 8192.
|