gguf quantized version of mochi (test pack for gguf-node)

screenshot

setup (once)

  • drag mochi-q3_k_m.gguf [4.31GB] to > ./ComfyUI/models/diffusion_models
  • drag t5xxl_fp16-q4_0.gguf [2.9GB] to > ./ComfyUI/models/text_encoders
  • drag mochi_vae_fp8_e4m3fn.safetensors [460MB] to > ./ComfyUI/models/vae

run it straight (no installation needed way)

  • run the .bat file in the main directory (assuming you are using the gguf-node pack below)
  • drag the workflow json file (below) to > your browser

workflow

review

  • revised workflow to bypass oom issue and around 50% faster with the new fp8_e4m3fn file
  • t5xxl works fine as text encoder; more quantized versions of t5xxl can be found here
  • gguf with pig architecture is working right away; welcome to test

reference

prompt test#

screenshot prompt: "a fox moving quickly in a beautiful winter scenery nature trees sunset tracking camera"

Prompt
a fox moving quickly in a beautiful winter scenery nature trees sunset tracking camera
Prompt
same prompt as 1st one <metadata inside>
Prompt
same prompt as 1st one; but with new workflow to bypass oom <metadata inside>
Downloads last month
114
GGUF
Model size
10B params
Architecture
mochi
Hardware compatibility
Log In to view the estimation

3-bit

4-bit

8-bit

16-bit

Inference Examples
Examples
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

This task can take several minutes

Model tree for calcuis/mochi

Quantized
(3)
this model