Update README.md
Browse files
README.md
CHANGED
@@ -21,6 +21,10 @@ license: other
|
|
21 |
|
22 |
These files are GGML format model files for [Gorilla LLM's Gorilla 7B](https://huggingface.co/gorilla-llm/gorilla-7b-hf-delta-v0).
|
23 |
|
|
|
|
|
|
|
|
|
24 |
GGML files are for CPU + GPU inference using [llama.cpp](https://github.com/ggerganov/llama.cpp) and libraries and UIs which support this format, such as:
|
25 |
* [text-generation-webui](https://github.com/oobabooga/text-generation-webui)
|
26 |
* [KoboldCpp](https://github.com/LostRuins/koboldcpp)
|
@@ -34,6 +38,13 @@ GGML files are for CPU + GPU inference using [llama.cpp](https://github.com/gger
|
|
34 |
* [2, 3, 4, 5, 6 and 8-bit GGML models for CPU+GPU inference](https://huggingface.co/TheBloke/gorilla-7B-GGML)
|
35 |
* [Unquantised fp16 model in pytorch format, for GPU inference and for further conversions](https://huggingface.co/TheBloke/gorilla-7B-fp16)
|
36 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
37 |
<!-- compatibility_ggml start -->
|
38 |
## Compatibility
|
39 |
|
|
|
21 |
|
22 |
These files are GGML format model files for [Gorilla LLM's Gorilla 7B](https://huggingface.co/gorilla-llm/gorilla-7b-hf-delta-v0).
|
23 |
|
24 |
+
**NOTE**: This is not a regular LLM. It is designed to allow LLMs to use tools by invoking APIs.
|
25 |
+
|
26 |
+
"Gorilla enables LLMs to use tools by invoking APIs. Given a natural language query, Gorilla can write a semantically- and syntactically- correct API to invoke. With Gorilla, we are the first to demonstrate how to use LLMs to invoke 1,600+ (and growing) API calls accurately while reducing hallucination. "
|
27 |
+
|
28 |
GGML files are for CPU + GPU inference using [llama.cpp](https://github.com/ggerganov/llama.cpp) and libraries and UIs which support this format, such as:
|
29 |
* [text-generation-webui](https://github.com/oobabooga/text-generation-webui)
|
30 |
* [KoboldCpp](https://github.com/LostRuins/koboldcpp)
|
|
|
38 |
* [2, 3, 4, 5, 6 and 8-bit GGML models for CPU+GPU inference](https://huggingface.co/TheBloke/gorilla-7B-GGML)
|
39 |
* [Unquantised fp16 model in pytorch format, for GPU inference and for further conversions](https://huggingface.co/TheBloke/gorilla-7B-fp16)
|
40 |
|
41 |
+
## Prompt template
|
42 |
+
|
43 |
+
```
|
44 |
+
###USER: find me an API to generate cute cat images
|
45 |
+
###ASSISTANT:
|
46 |
+
```
|
47 |
+
|
48 |
<!-- compatibility_ggml start -->
|
49 |
## Compatibility
|
50 |
|