Update README.md
Browse fileshttps://github.com/turboderp-org/exllamav3/commit/c44e56c73b2c67eee087c7195c9093520494d3bf
https://github.com/turboderp-org/exllamav2/commit/de19cbcc599353d5aee1fec8c1ce2806f890baca
https://huggingface.co/Disya/GLM4-9B-Neon-v2-exl2-5.5bpw-h8
exl3 and exl2 supports
README.md
CHANGED
@@ -58,7 +58,7 @@ To run GGUFs correctly, you need the most recent version of KoboldCPP, and to pa
|
|
58 |
|
59 |
Thanks to DaringDuck and tofumagnate for info how to apply this fix.
|
60 |
|
61 |
-
To run this model on vLLM, you'll need to build it from source from the git repo, GLM4 support haven't reached release yet.
|
62 |
|
63 |
---
|
64 |
|
|
|
58 |
|
59 |
Thanks to DaringDuck and tofumagnate for info how to apply this fix.
|
60 |
|
61 |
+
To run this model on vLLM, you'll need to build it from source from the git repo, GLM4 support haven't reached release yet.
|
62 |
|
63 |
---
|
64 |
|