Update README.md
Browse files
README.md
CHANGED
@@ -5,12 +5,17 @@ tags:
|
|
5 |
- merge
|
6 |
- Roleplay
|
7 |
---
|
8 |
-
# Experimental model
|
9 |
A Experimental MoE Model that custom for all-rounded Roleplay. Well understand Character Card and high logic.
|
|
|
10 |
# It's ridiculous that I can run this original version in 4bit, but can't run in GGUF version. I hope someone could help me fix the GGUF version.
|
|
|
11 |
This is the following error line that I got when try loading GGUF version of this model:
|
12 |
|
13 |
``GGML_ASSERT: /home/runner/work/llama-cpp-python-cuBLAS-wheels/llama-cpp-python-cuBLAS-wheels/vendor/llama.cpp/ggml-cuda.cu:8431: (ncols & (ncols - 1)) == 0``
|
14 |
|
15 |
Have try from Q2 to fp16, no hope. 😥
|
|
|
|
|
|
|
16 |
# Is this model good? Want more dicussion? Let's me know in community tab! ヾ(≧▽≦*)o
|
|
|
5 |
- merge
|
6 |
- Roleplay
|
7 |
---
|
8 |
+
# Experimental 3x7B model
|
9 |
A Experimental MoE Model that custom for all-rounded Roleplay. Well understand Character Card and high logic.
|
10 |
+
|
11 |
# It's ridiculous that I can run this original version in 4bit, but can't run in GGUF version. I hope someone could help me fix the GGUF version.
|
12 |
+
|
13 |
This is the following error line that I got when try loading GGUF version of this model:
|
14 |
|
15 |
``GGML_ASSERT: /home/runner/work/llama-cpp-python-cuBLAS-wheels/llama-cpp-python-cuBLAS-wheels/vendor/llama.cpp/ggml-cuda.cu:8431: (ncols & (ncols - 1)) == 0``
|
16 |
|
17 |
Have try from Q2 to fp16, no hope. 😥
|
18 |
+
|
19 |
+
Link here: https://huggingface.co/Alsebay/HyouKan-GGUF
|
20 |
+
|
21 |
# Is this model good? Want more dicussion? Let's me know in community tab! ヾ(≧▽≦*)o
|