Create README.md
Browse files
README.md
ADDED
@@ -0,0 +1,20 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
language:
|
3 |
+
- fr
|
4 |
+
- it
|
5 |
+
- de
|
6 |
+
- es
|
7 |
+
- en
|
8 |
+
license: apache-2.0
|
9 |
+
inference:
|
10 |
+
parameters:
|
11 |
+
temperature: 0.5
|
12 |
+
---
|
13 |
+
|
14 |
+
This is [mistralai/Mixtral-8x7B-Instruct-v0.1](https://huggingface.co/mistralai/Mixtral-8x7B-Instruct-v0.1), converted to GGUF and quantized to q8_0. Both the model and the embedding/output tensors are q8_0.
|
15 |
+
|
16 |
+
The model is split using the `llama.cpp/llama-gguf-split` cli utility into shards no larger than 1GB. The purpose of this is to make it less painful to resume downloading if interrupted.
|
17 |
+
|
18 |
+
This is uploaded pretty much just as a personal backup. Mixtral Instruct is one of my favorite models.
|
19 |
+
|
20 |
+
All operations are done with `llama.cpp` commit [`8cd1bcfd3fc9f2b5cbafd7fb7581b3278acec25fz`](https://github.com/ggerganov/llama.cpp/tree/8cd1bcfd3fc9f2b5cbafd7fb7581b3278acec25fz) (2024-08-11).
|