FLUX schnell Quantized Models

This repo contains quantized versions of the FLUX schnell transformer for use in InvokeAI.

Contents:

  • transformer/base/ - Transformer in bfloat16 copied from here
  • transformer/bnb_nf4/ - Transformer quantized to bitsandbytes NF4 format using this script
Downloads last month

-

Downloads are not tracked for this model. How to track
Inference API
Unable to determine this model's library. Check the docs .