Update README.md
Browse files
README.md
CHANGED
@@ -59,7 +59,7 @@ To load the 4‑bit quantized HiDream I1 model in ComfyUI using the custom loade
|
|
59 |
## Quick Q&A
|
60 |
|
61 |
**1. What’s the difference from the official NF4?**
|
62 |
-
The official RediDream NF4 gguf version is actually a gguf quantized version of Q4_K_M, not a real bnb-nf4. So the official version is slightly larger than ours (10.7GB vs 9.6GB), and the effect is better. Our NF4 model needs about 25+ steps to reach the level of the official NF4, and each step is 30% slower on average. But our
|
63 |
|
64 |
**2. How’s the performance?**
|
65 |
The quantized model requires more conservative settings (e.g. skipping layers 2–6, 15–25 steps, CFG 1.0) to produce high-quality images, so generation may be a bit slower(single step 1.5s vs 1.1 s on RTX 4080).
|
|
|
59 |
## Quick Q&A
|
60 |
|
61 |
**1. What’s the difference from the official NF4?**
|
62 |
+
The official RediDream NF4 gguf version is actually a gguf quantized version of Q4_K_M, not a real bnb-nf4. So the official version is slightly larger than ours (10.7GB vs 9.6GB), and the effect is better. Our NF4 model needs about 25+ steps to reach the level of the official NF4, and each step is 30% slower on average. But our VRAM usage is also slightly smaller than the official one (11.6GB vs 13.3 GB). Overall, the level of our model is almost the same as the Q4_0 quantization.
|
63 |
|
64 |
**2. How’s the performance?**
|
65 |
The quantized model requires more conservative settings (e.g. skipping layers 2–6, 15–25 steps, CFG 1.0) to produce high-quality images, so generation may be a bit slower(single step 1.5s vs 1.1 s on RTX 4080).
|