Update README.md
Browse files
README.md
CHANGED
@@ -34,7 +34,7 @@ This approach reduces memory usage during quantization to ~40 GB and is **entire
|
|
34 |
|
35 |
## Performance
|
36 |
|
37 |
-
- The quantized models perform **similarly** to the
|
38 |
- No formal benchmark or deep evaluation has been conducted.
|
39 |
|
40 |
## Files
|
|
|
34 |
|
35 |
## Performance
|
36 |
|
37 |
+
- The quantized models perform **similarly** to the Q4_0 quantized version.
|
38 |
- No formal benchmark or deep evaluation has been conducted.
|
39 |
|
40 |
## Files
|