Add llama.cpp to the examples
Browse filesLet's add llama.cpp to the examples, considering 2 of the other applications mentioned are powered by the llama.cpp library (ggml).
https://github.com/ggml-org/ggml
https://github.com/ggml-org/llama.cpp
README.md
CHANGED
@@ -394,6 +394,23 @@ docker run -it --rm --pull=always \
|
|
394 |
Click “see advanced setting” on the second line.
|
395 |
In the new tab, toggle advanced to on. Set the custom model to be mistral/devstralq4_k_m and Base URL the api address we get from the last step in LM Studio. Set API Key to dummy. Click save changes.
|
396 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
397 |
|
398 |
### Ollama
|
399 |
|
|
|
394 |
Click “see advanced setting” on the second line.
|
395 |
In the new tab, toggle advanced to on. Set the custom model to be mistral/devstralq4_k_m and Base URL the api address we get from the last step in LM Studio. Set API Key to dummy. Click save changes.
|
396 |
|
397 |
+
### llama.cpp
|
398 |
+
|
399 |
+
Download the weights from huggingface:
|
400 |
+
|
401 |
+
```
|
402 |
+
pip install -U "huggingface_hub[cli]"
|
403 |
+
huggingface-cli download \
|
404 |
+
"mistralai/Devstral-Small-2505_gguf" \
|
405 |
+
--include "devstralQ4_K_M.gguf" \
|
406 |
+
--local-dir "mistralai/Devstral-Small-2505_gguf/"
|
407 |
+
```
|
408 |
+
|
409 |
+
Then run Devstral using the llama.cpp CLI.
|
410 |
+
|
411 |
+
```bash
|
412 |
+
./llama-cli -m Devstral-Small-2505_gguf/devstralQ4_K_M.gguf -cnv
|
413 |
+
```
|
414 |
|
415 |
### Ollama
|
416 |
|