b-a-s-e-d commited on
Commit
184a0a7
·
verified ·
1 Parent(s): 887d1e7

Add llama.cpp to the examples

Browse files

Let's add llama.cpp to the examples, considering 2 of the other applications mentioned are powered by the llama.cpp library (ggml).

https://github.com/ggml-org/ggml
https://github.com/ggml-org/llama.cpp

Files changed (1) hide show
  1. README.md +17 -0
README.md CHANGED
@@ -394,6 +394,23 @@ docker run -it --rm --pull=always \
394
  Click “see advanced setting” on the second line.
395
  In the new tab, toggle advanced to on. Set the custom model to be mistral/devstralq4_k_m and Base URL the api address we get from the last step in LM Studio. Set API Key to dummy. Click save changes.
396
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
397
 
398
  ### Ollama
399
 
 
394
  Click “see advanced setting” on the second line.
395
  In the new tab, toggle advanced to on. Set the custom model to be mistral/devstralq4_k_m and Base URL the api address we get from the last step in LM Studio. Set API Key to dummy. Click save changes.
396
 
397
+ ### llama.cpp
398
+
399
+ Download the weights from huggingface:
400
+
401
+ ```
402
+ pip install -U "huggingface_hub[cli]"
403
+ huggingface-cli download \
404
+ "mistralai/Devstral-Small-2505_gguf" \
405
+ --include "devstralQ4_K_M.gguf" \
406
+ --local-dir "mistralai/Devstral-Small-2505_gguf/"
407
+ ```
408
+
409
+ Then run Devstral using the llama.cpp CLI.
410
+
411
+ ```bash
412
+ ./llama-cli -m Devstral-Small-2505_gguf/devstralQ4_K_M.gguf -cnv
413
+ ```
414
 
415
  ### Ollama
416