Add library_name, pipeline_tag and set inference to true
#15
by
nielsr
HF Staff
- opened
README.md
CHANGED
@@ -1,13 +1,15 @@
|
|
1 |
---
|
2 |
-
license: apache-2.0
|
3 |
language:
|
4 |
-
|
|
|
|
|
|
|
5 |
tags:
|
6 |
-
|
7 |
-
|
8 |
-
|
9 |
-
|
10 |
-
inference:
|
11 |
---
|
12 |
|
13 |
# CogVideoX-2B
|
@@ -180,7 +182,7 @@ pipe.vae.enable_tiling()
|
|
180 |
+ The 2B model is trained with `FP16` precision, and the 5B model is trained with `BF16` precision. We recommend using
|
181 |
the precision the model was trained with for inference.
|
182 |
+ [PytorchAO](https://github.com/pytorch/ao) and [Optimum-quanto](https://github.com/huggingface/optimum-quanto/) can be
|
183 |
-
used to quantize the text encoder,
|
184 |
it possible to run the model on a free T4 Colab or GPUs with smaller VRAM! It is also worth noting that TorchAO
|
185 |
quantization is fully compatible with `torch.compile`, which can significantly improve inference speed. `FP8`
|
186 |
precision must be used on devices with `NVIDIA H100` or above, which requires installing
|
|
|
1 |
---
|
|
|
2 |
language:
|
3 |
+
- en
|
4 |
+
license: apache-2.0
|
5 |
+
library_name: diffusers
|
6 |
+
pipeline_tag: text-to-video
|
7 |
tags:
|
8 |
+
- cogvideox
|
9 |
+
- video-generation
|
10 |
+
- thudm
|
11 |
+
- text-to-video
|
12 |
+
inference: true
|
13 |
---
|
14 |
|
15 |
# CogVideoX-2B
|
|
|
182 |
+ The 2B model is trained with `FP16` precision, and the 5B model is trained with `BF16` precision. We recommend using
|
183 |
the precision the model was trained with for inference.
|
184 |
+ [PytorchAO](https://github.com/pytorch/ao) and [Optimum-quanto](https://github.com/huggingface/optimum-quanto/) can be
|
185 |
+
used to quantize the text encoder, transformer, and VAE modules to reduce CogVideoX's memory requirements. This makes
|
186 |
it possible to run the model on a free T4 Colab or GPUs with smaller VRAM! It is also worth noting that TorchAO
|
187 |
quantization is fully compatible with `torch.compile`, which can significantly improve inference speed. `FP8`
|
188 |
precision must be used on devices with `NVIDIA H100` or above, which requires installing
|