Update README.md
Browse files
README.md
CHANGED
@@ -110,9 +110,6 @@ The following clients/libraries will automatically download models for you, prov
|
|
110 |
- LoLLMS Web UI
|
111 |
- Faraday.dev
|
112 |
|
113 |
-
### In `text-generation-webui`
|
114 |
-
|
115 |
-
Under Download Model, you can enter the model repo: TheBloke/CodeLlama-13B-Instruct-GGUF and below it, a specific filename to download, such as: codellama-13b-instruct.q4_K_M.gguf.
|
116 |
|
117 |
Then click Download.
|
118 |
|
@@ -149,9 +146,6 @@ pip3 install hf_transfer
|
|
149 |
|
150 |
And set environment variable `HF_HUB_ENABLE_HF_TRANSFER` to `1`:
|
151 |
|
152 |
-
```shell
|
153 |
-
HUGGINGFACE_HUB_ENABLE_HF_TRANSFER=1 huggingface-cli download TheBloke/CodeLlama-13B-Instruct-GGUF codellama-13b-instruct.q4_K_M.gguf --local-dir . --local-dir-use-symlinks False
|
154 |
-
```
|
155 |
|
156 |
Windows CLI users: Use `set HUGGINGFACE_HUB_ENABLE_HF_TRANSFER=1` before running the download command.
|
157 |
</details>
|
|
|
110 |
- LoLLMS Web UI
|
111 |
- Faraday.dev
|
112 |
|
|
|
|
|
|
|
113 |
|
114 |
Then click Download.
|
115 |
|
|
|
146 |
|
147 |
And set environment variable `HF_HUB_ENABLE_HF_TRANSFER` to `1`:
|
148 |
|
|
|
|
|
|
|
149 |
|
150 |
Windows CLI users: Use `set HUGGINGFACE_HUB_ENABLE_HF_TRANSFER=1` before running the download command.
|
151 |
</details>
|