Text Generation
Transformers
Safetensors
PyTorch
English
llama
facebook
meta
llama-2
text-generation-inference
uukuguy commited on
Commit
6ff9365
1 Parent(s): 5341819

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -0
README.md CHANGED
@@ -27,6 +27,12 @@ tags:
27
  ---
28
 
29
  <p><h1> speechless-llama2-13b:v1.1 </h1></p>
 
 
 
 
 
 
30
  speechless-llama2-13b:v1.1 is a merge of Open-Orca/OpenOrca-Platypus2-13B and WizardLM/WizardLM-13B-V1.2.
31
 
32
  | Metric | Value |
 
27
  ---
28
 
29
  <p><h1> speechless-llama2-13b:v1.1 </h1></p>
30
+
31
+ * [GPTQ models for GPU inference, with multiple quantisation parameter options.](https://huggingface.co/TheBloke/Speechless-Llama2-13B-GPTQ)
32
+ * [2, 3, 4, 5, 6 and 8-bit GGUF models for CPU+GPU inference](https://huggingface.co/TheBloke/Speechless-Llama2-13B-GGUF)
33
+ * [2, 3, 4, 5, 6 and 8-bit GGML models for CPU+GPU inference (deprecated)](https://huggingface.co/TheBloke/Speechless-Llama2-13B-GGML)
34
+
35
+
36
  speechless-llama2-13b:v1.1 is a merge of Open-Orca/OpenOrca-Platypus2-13B and WizardLM/WizardLM-13B-V1.2.
37
 
38
  | Metric | Value |