Initial GPTQ model commit
Browse files
README.md
CHANGED
@@ -40,7 +40,7 @@ GGML versions are not yet provided, as there is not yet support for SuperHOT in
|
|
40 |
## Repositories available
|
41 |
|
42 |
* [4-bit GPTQ models for GPU inference](https://huggingface.co/TheBloke/Chronos-13B-SuperHOT-8K-GPTQ)
|
43 |
-
* [2, 3, 4, 5, 6 and 8-bit GGML models for CPU inference](https://huggingface.co/
|
44 |
* [Unquantised SuperHOT fp16 model in pytorch format, for GPU inference and for further conversions](https://huggingface.co/TheBloke/Chronos-13B-SuperHOT-8K-fp16)
|
45 |
* [Unquantised base fp16 model in pytorch format, for GPU inference and for further conversions](https://huggingface.co/elinas/chronos-13b)
|
46 |
|
|
|
40 |
## Repositories available
|
41 |
|
42 |
* [4-bit GPTQ models for GPU inference](https://huggingface.co/TheBloke/Chronos-13B-SuperHOT-8K-GPTQ)
|
43 |
+
* [2, 3, 4, 5, 6 and 8-bit GGML models for CPU inference](https://huggingface.co/TheBloke/Chronos-13B-SuperHOT-8K-GGML)
|
44 |
* [Unquantised SuperHOT fp16 model in pytorch format, for GPU inference and for further conversions](https://huggingface.co/TheBloke/Chronos-13B-SuperHOT-8K-fp16)
|
45 |
* [Unquantised base fp16 model in pytorch format, for GPU inference and for further conversions](https://huggingface.co/elinas/chronos-13b)
|
46 |
|