Upload folder using huggingface_hub
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- main/composable_stable_diffusion.py +1 -1
- main/imagic_stable_diffusion.py +1 -1
- main/img2img_inpainting.py +1 -1
- main/interpolate_stable_diffusion.py +1 -1
- main/lpw_stable_diffusion.py +2 -2
- main/lpw_stable_diffusion_onnx.py +2 -2
- main/lpw_stable_diffusion_xl.py +1 -1
- main/multilingual_stable_diffusion.py +1 -1
- main/pipeline_controlnet_xl_kolors.py +1 -1
- main/pipeline_controlnet_xl_kolors_img2img.py +1 -1
- main/pipeline_controlnet_xl_kolors_inpaint.py +1 -1
- main/pipeline_demofusion_sdxl.py +1 -1
- main/pipeline_faithdiff_stable_diffusion_xl.py +1 -1
- main/pipeline_flux_differential_img2img.py +2 -2
- main/pipeline_flux_kontext_multiple_images.py +1 -1
- main/pipeline_flux_rf_inversion.py +1 -1
- main/pipeline_flux_semantic_guidance.py +1 -1
- main/pipeline_flux_with_cfg.py +1 -1
- main/pipeline_kolors_differential_img2img.py +1 -1
- main/pipeline_kolors_inpainting.py +1 -1
- main/pipeline_prompt2prompt.py +1 -1
- main/pipeline_sdxl_style_aligned.py +1 -1
- main/pipeline_stable_diffusion_3_differential_img2img.py +1 -1
- main/pipeline_stable_diffusion_3_instruct_pix2pix.py +1 -1
- main/pipeline_stable_diffusion_xl_attentive_eraser.py +1 -1
- main/pipeline_stable_diffusion_xl_controlnet_adapter.py +1 -1
- main/pipeline_stable_diffusion_xl_controlnet_adapter_inpaint.py +1 -1
- main/pipeline_stable_diffusion_xl_differential_img2img.py +1 -1
- main/pipeline_stable_diffusion_xl_ipex.py +1 -1
- main/pipeline_stg_cogvideox.py +1 -1
- main/pipeline_stg_ltx.py +1 -1
- main/pipeline_stg_ltx_image2video.py +1 -1
- main/pipeline_stg_mochi.py +1 -1
- main/pipeline_zero1to3.py +1 -1
- main/rerender_a_video.py +1 -1
- main/run_onnx_controlnet.py +1 -1
- main/run_tensorrt_controlnet.py +1 -1
- main/sd_text2img_k_diffusion.py +1 -1
- main/seed_resize_stable_diffusion.py +1 -1
- main/stable_diffusion_comparison.py +1 -1
- main/stable_diffusion_controlnet_img2img.py +1 -1
- main/stable_diffusion_controlnet_inpaint.py +1 -1
- main/stable_diffusion_controlnet_inpaint_img2img.py +1 -1
- main/stable_diffusion_controlnet_reference.py +1 -1
- main/stable_diffusion_ipex.py +1 -1
- main/stable_diffusion_reference.py +1 -1
- main/stable_diffusion_repaint.py +1 -1
- main/stable_diffusion_xl_reference.py +1 -1
- main/text_inpainting.py +1 -1
- main/tiled_upscaling.py +1 -1
main/composable_stable_diffusion.py
CHANGED
|
@@ -398,7 +398,7 @@ class ComposableStableDiffusionPipeline(DiffusionPipeline, StableDiffusionMixin)
|
|
| 398 |
latents (`torch.Tensor`, *optional*):
|
| 399 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 400 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 401 |
-
tensor will
|
| 402 |
output_type (`str`, *optional*, defaults to `"pil"`):
|
| 403 |
The output format of the generate image. Choose between
|
| 404 |
[PIL](https://pillow.readthedocs.io/en/stable/): `PIL.Image.Image` or `np.array`.
|
|
|
|
| 398 |
latents (`torch.Tensor`, *optional*):
|
| 399 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 400 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 401 |
+
tensor will be generated by sampling using the supplied random `generator`.
|
| 402 |
output_type (`str`, *optional*, defaults to `"pil"`):
|
| 403 |
The output format of the generate image. Choose between
|
| 404 |
[PIL](https://pillow.readthedocs.io/en/stable/): `PIL.Image.Image` or `np.array`.
|
main/imagic_stable_diffusion.py
CHANGED
|
@@ -147,7 +147,7 @@ class ImagicStableDiffusionPipeline(DiffusionPipeline, StableDiffusionMixin):
|
|
| 147 |
latents (`torch.Tensor`, *optional*):
|
| 148 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 149 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 150 |
-
tensor will
|
| 151 |
output_type (`str`, *optional*, defaults to `"pil"`):
|
| 152 |
The output format of the generate image. Choose between
|
| 153 |
[PIL](https://pillow.readthedocs.io/en/stable/): `PIL.Image.Image` or `nd.array`.
|
|
|
|
| 147 |
latents (`torch.Tensor`, *optional*):
|
| 148 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 149 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 150 |
+
tensor will be generated by sampling using the supplied random `generator`.
|
| 151 |
output_type (`str`, *optional*, defaults to `"pil"`):
|
| 152 |
The output format of the generate image. Choose between
|
| 153 |
[PIL](https://pillow.readthedocs.io/en/stable/): `PIL.Image.Image` or `nd.array`.
|
main/img2img_inpainting.py
CHANGED
|
@@ -197,7 +197,7 @@ class ImageToImageInpaintingPipeline(DiffusionPipeline):
|
|
| 197 |
latents (`torch.Tensor`, *optional*):
|
| 198 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 199 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 200 |
-
tensor will
|
| 201 |
output_type (`str`, *optional*, defaults to `"pil"`):
|
| 202 |
The output format of the generate image. Choose between
|
| 203 |
[PIL](https://pillow.readthedocs.io/en/stable/): `PIL.Image.Image` or `np.array`.
|
|
|
|
| 197 |
latents (`torch.Tensor`, *optional*):
|
| 198 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 199 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 200 |
+
tensor will be generated by sampling using the supplied random `generator`.
|
| 201 |
output_type (`str`, *optional*, defaults to `"pil"`):
|
| 202 |
The output format of the generate image. Choose between
|
| 203 |
[PIL](https://pillow.readthedocs.io/en/stable/): `PIL.Image.Image` or `np.array`.
|
main/interpolate_stable_diffusion.py
CHANGED
|
@@ -173,7 +173,7 @@ class StableDiffusionWalkPipeline(DiffusionPipeline, StableDiffusionMixin):
|
|
| 173 |
latents (`torch.Tensor`, *optional*):
|
| 174 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 175 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 176 |
-
tensor will
|
| 177 |
output_type (`str`, *optional*, defaults to `"pil"`):
|
| 178 |
The output format of the generate image. Choose between
|
| 179 |
[PIL](https://pillow.readthedocs.io/en/stable/): `PIL.Image.Image` or `np.array`.
|
|
|
|
| 173 |
latents (`torch.Tensor`, *optional*):
|
| 174 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 175 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 176 |
+
tensor will be generated by sampling using the supplied random `generator`.
|
| 177 |
output_type (`str`, *optional*, defaults to `"pil"`):
|
| 178 |
The output format of the generate image. Choose between
|
| 179 |
[PIL](https://pillow.readthedocs.io/en/stable/): `PIL.Image.Image` or `np.array`.
|
main/lpw_stable_diffusion.py
CHANGED
|
@@ -888,7 +888,7 @@ class StableDiffusionLongPromptWeightingPipeline(
|
|
| 888 |
latents (`torch.Tensor`, *optional*):
|
| 889 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 890 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 891 |
-
tensor will
|
| 892 |
prompt_embeds (`torch.Tensor`, *optional*):
|
| 893 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 894 |
provided, text embeddings will be generated from `prompt` input argument.
|
|
@@ -1131,7 +1131,7 @@ class StableDiffusionLongPromptWeightingPipeline(
|
|
| 1131 |
latents (`torch.Tensor`, *optional*):
|
| 1132 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 1133 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 1134 |
-
tensor will
|
| 1135 |
prompt_embeds (`torch.Tensor`, *optional*):
|
| 1136 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 1137 |
provided, text embeddings will be generated from `prompt` input argument.
|
|
|
|
| 888 |
latents (`torch.Tensor`, *optional*):
|
| 889 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 890 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 891 |
+
tensor will be generated by sampling using the supplied random `generator`.
|
| 892 |
prompt_embeds (`torch.Tensor`, *optional*):
|
| 893 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 894 |
provided, text embeddings will be generated from `prompt` input argument.
|
|
|
|
| 1131 |
latents (`torch.Tensor`, *optional*):
|
| 1132 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 1133 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 1134 |
+
tensor will be generated by sampling using the supplied random `generator`.
|
| 1135 |
prompt_embeds (`torch.Tensor`, *optional*):
|
| 1136 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 1137 |
provided, text embeddings will be generated from `prompt` input argument.
|
main/lpw_stable_diffusion_onnx.py
CHANGED
|
@@ -721,7 +721,7 @@ class OnnxStableDiffusionLongPromptWeightingPipeline(OnnxStableDiffusionPipeline
|
|
| 721 |
latents (`np.ndarray`, *optional*):
|
| 722 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 723 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 724 |
-
tensor will
|
| 725 |
max_embeddings_multiples (`int`, *optional*, defaults to `3`):
|
| 726 |
The max multiple length of prompt embeddings compared to the max output length of text encoder.
|
| 727 |
output_type (`str`, *optional*, defaults to `"pil"`):
|
|
@@ -918,7 +918,7 @@ class OnnxStableDiffusionLongPromptWeightingPipeline(OnnxStableDiffusionPipeline
|
|
| 918 |
latents (`np.ndarray`, *optional*):
|
| 919 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 920 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 921 |
-
tensor will
|
| 922 |
max_embeddings_multiples (`int`, *optional*, defaults to `3`):
|
| 923 |
The max multiple length of prompt embeddings compared to the max output length of text encoder.
|
| 924 |
output_type (`str`, *optional*, defaults to `"pil"`):
|
|
|
|
| 721 |
latents (`np.ndarray`, *optional*):
|
| 722 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 723 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 724 |
+
tensor will be generated by sampling using the supplied random `generator`.
|
| 725 |
max_embeddings_multiples (`int`, *optional*, defaults to `3`):
|
| 726 |
The max multiple length of prompt embeddings compared to the max output length of text encoder.
|
| 727 |
output_type (`str`, *optional*, defaults to `"pil"`):
|
|
|
|
| 918 |
latents (`np.ndarray`, *optional*):
|
| 919 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 920 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 921 |
+
tensor will be generated by sampling using the supplied random `generator`.
|
| 922 |
max_embeddings_multiples (`int`, *optional*, defaults to `3`):
|
| 923 |
The max multiple length of prompt embeddings compared to the max output length of text encoder.
|
| 924 |
output_type (`str`, *optional*, defaults to `"pil"`):
|
main/lpw_stable_diffusion_xl.py
CHANGED
|
@@ -1519,7 +1519,7 @@ class SDXLLongPromptWeightingPipeline(
|
|
| 1519 |
latents (`torch.Tensor`, *optional*):
|
| 1520 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 1521 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 1522 |
-
tensor will
|
| 1523 |
ip_adapter_image: (`PipelineImageInput`, *optional*):
|
| 1524 |
Optional image input to work with IP Adapters.
|
| 1525 |
prompt_embeds (`torch.Tensor`, *optional*):
|
|
|
|
| 1519 |
latents (`torch.Tensor`, *optional*):
|
| 1520 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 1521 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 1522 |
+
tensor will be generated by sampling using the supplied random `generator`.
|
| 1523 |
ip_adapter_image: (`PipelineImageInput`, *optional*):
|
| 1524 |
Optional image input to work with IP Adapters.
|
| 1525 |
prompt_embeds (`torch.Tensor`, *optional*):
|
main/multilingual_stable_diffusion.py
CHANGED
|
@@ -187,7 +187,7 @@ class MultilingualStableDiffusion(DiffusionPipeline, StableDiffusionMixin):
|
|
| 187 |
latents (`torch.Tensor`, *optional*):
|
| 188 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 189 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 190 |
-
tensor will
|
| 191 |
output_type (`str`, *optional*, defaults to `"pil"`):
|
| 192 |
The output format of the generate image. Choose between
|
| 193 |
[PIL](https://pillow.readthedocs.io/en/stable/): `PIL.Image.Image` or `np.array`.
|
|
|
|
| 187 |
latents (`torch.Tensor`, *optional*):
|
| 188 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 189 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 190 |
+
tensor will be generated by sampling using the supplied random `generator`.
|
| 191 |
output_type (`str`, *optional*, defaults to `"pil"`):
|
| 192 |
The output format of the generate image. Choose between
|
| 193 |
[PIL](https://pillow.readthedocs.io/en/stable/): `PIL.Image.Image` or `np.array`.
|
main/pipeline_controlnet_xl_kolors.py
CHANGED
|
@@ -888,7 +888,7 @@ class KolorsControlNetPipeline(
|
|
| 888 |
latents (`torch.Tensor`, *optional*):
|
| 889 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 890 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 891 |
-
tensor will
|
| 892 |
prompt_embeds (`torch.Tensor`, *optional*):
|
| 893 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 894 |
provided, text embeddings will be generated from `prompt` input argument.
|
|
|
|
| 888 |
latents (`torch.Tensor`, *optional*):
|
| 889 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 890 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 891 |
+
tensor will be generated by sampling using the supplied random `generator`.
|
| 892 |
prompt_embeds (`torch.Tensor`, *optional*):
|
| 893 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 894 |
provided, text embeddings will be generated from `prompt` input argument.
|
main/pipeline_controlnet_xl_kolors_img2img.py
CHANGED
|
@@ -1066,7 +1066,7 @@ class KolorsControlNetImg2ImgPipeline(
|
|
| 1066 |
latents (`torch.Tensor`, *optional*):
|
| 1067 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 1068 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 1069 |
-
tensor will
|
| 1070 |
prompt_embeds (`torch.Tensor`, *optional*):
|
| 1071 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 1072 |
provided, text embeddings will be generated from `prompt` input argument.
|
|
|
|
| 1066 |
latents (`torch.Tensor`, *optional*):
|
| 1067 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 1068 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 1069 |
+
tensor will be generated by sampling using the supplied random `generator`.
|
| 1070 |
prompt_embeds (`torch.Tensor`, *optional*):
|
| 1071 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 1072 |
provided, text embeddings will be generated from `prompt` input argument.
|
main/pipeline_controlnet_xl_kolors_inpaint.py
CHANGED
|
@@ -1298,7 +1298,7 @@ class KolorsControlNetInpaintPipeline(
|
|
| 1298 |
latents (`torch.Tensor`, *optional*):
|
| 1299 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 1300 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 1301 |
-
tensor will
|
| 1302 |
output_type (`str`, *optional*, defaults to `"pil"`):
|
| 1303 |
The output format of the generate image. Choose between
|
| 1304 |
[PIL](https://pillow.readthedocs.io/en/stable/): `PIL.Image.Image` or `np.array`.
|
|
|
|
| 1298 |
latents (`torch.Tensor`, *optional*):
|
| 1299 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 1300 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 1301 |
+
tensor will be generated by sampling using the supplied random `generator`.
|
| 1302 |
output_type (`str`, *optional*, defaults to `"pil"`):
|
| 1303 |
The output format of the generate image. Choose between
|
| 1304 |
[PIL](https://pillow.readthedocs.io/en/stable/): `PIL.Image.Image` or `np.array`.
|
main/pipeline_demofusion_sdxl.py
CHANGED
|
@@ -724,7 +724,7 @@ class DemoFusionSDXLPipeline(
|
|
| 724 |
latents (`torch.Tensor`, *optional*):
|
| 725 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 726 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 727 |
-
tensor will
|
| 728 |
prompt_embeds (`torch.Tensor`, *optional*):
|
| 729 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 730 |
provided, text embeddings will be generated from `prompt` input argument.
|
|
|
|
| 724 |
latents (`torch.Tensor`, *optional*):
|
| 725 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 726 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 727 |
+
tensor will be generated by sampling using the supplied random `generator`.
|
| 728 |
prompt_embeds (`torch.Tensor`, *optional*):
|
| 729 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 730 |
provided, text embeddings will be generated from `prompt` input argument.
|
main/pipeline_faithdiff_stable_diffusion_xl.py
CHANGED
|
@@ -1906,7 +1906,7 @@ class FaithDiffStableDiffusionXLPipeline(
|
|
| 1906 |
latents (`torch.FloatTensor`, *optional*):
|
| 1907 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 1908 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 1909 |
-
tensor will
|
| 1910 |
prompt_embeds (`torch.FloatTensor`, *optional*):
|
| 1911 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 1912 |
provided, text embeddings will be generated from `prompt` input argument.
|
|
|
|
| 1906 |
latents (`torch.FloatTensor`, *optional*):
|
| 1907 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 1908 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 1909 |
+
tensor will be generated by sampling using the supplied random `generator`.
|
| 1910 |
prompt_embeds (`torch.FloatTensor`, *optional*):
|
| 1911 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 1912 |
provided, text embeddings will be generated from `prompt` input argument.
|
main/pipeline_flux_differential_img2img.py
CHANGED
|
@@ -730,7 +730,7 @@ class FluxDifferentialImg2ImgPipeline(DiffusionPipeline, FluxLoraLoaderMixin):
|
|
| 730 |
1)`, or `(H, W)`.
|
| 731 |
mask_image_latent (`torch.Tensor`, `List[torch.Tensor]`):
|
| 732 |
`Tensor` representing an image batch to mask `image` generated by VAE. If not provided, the mask
|
| 733 |
-
latents tensor will
|
| 734 |
height (`int`, *optional*, defaults to self.unet.config.sample_size * self.vae_scale_factor):
|
| 735 |
The height in pixels of the generated image. This is set to 1024 by default for the best results.
|
| 736 |
width (`int`, *optional*, defaults to self.unet.config.sample_size * self.vae_scale_factor):
|
|
@@ -769,7 +769,7 @@ class FluxDifferentialImg2ImgPipeline(DiffusionPipeline, FluxLoraLoaderMixin):
|
|
| 769 |
latents (`torch.FloatTensor`, *optional*):
|
| 770 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 771 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 772 |
-
tensor will
|
| 773 |
prompt_embeds (`torch.FloatTensor`, *optional*):
|
| 774 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 775 |
provided, text embeddings will be generated from `prompt` input argument.
|
|
|
|
| 730 |
1)`, or `(H, W)`.
|
| 731 |
mask_image_latent (`torch.Tensor`, `List[torch.Tensor]`):
|
| 732 |
`Tensor` representing an image batch to mask `image` generated by VAE. If not provided, the mask
|
| 733 |
+
latents tensor will be generated by `mask_image`.
|
| 734 |
height (`int`, *optional*, defaults to self.unet.config.sample_size * self.vae_scale_factor):
|
| 735 |
The height in pixels of the generated image. This is set to 1024 by default for the best results.
|
| 736 |
width (`int`, *optional*, defaults to self.unet.config.sample_size * self.vae_scale_factor):
|
|
|
|
| 769 |
latents (`torch.FloatTensor`, *optional*):
|
| 770 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 771 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 772 |
+
tensor will be generated by sampling using the supplied random `generator`.
|
| 773 |
prompt_embeds (`torch.FloatTensor`, *optional*):
|
| 774 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 775 |
provided, text embeddings will be generated from `prompt` input argument.
|
main/pipeline_flux_kontext_multiple_images.py
CHANGED
|
@@ -885,7 +885,7 @@ class FluxKontextPipeline(
|
|
| 885 |
latents (`torch.FloatTensor`, *optional*):
|
| 886 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 887 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 888 |
-
tensor will
|
| 889 |
prompt_embeds (`torch.FloatTensor`, *optional*):
|
| 890 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 891 |
provided, text embeddings will be generated from `prompt` input argument.
|
|
|
|
| 885 |
latents (`torch.FloatTensor`, *optional*):
|
| 886 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 887 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 888 |
+
tensor will be generated by sampling using the supplied random `generator`.
|
| 889 |
prompt_embeds (`torch.FloatTensor`, *optional*):
|
| 890 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 891 |
provided, text embeddings will be generated from `prompt` input argument.
|
main/pipeline_flux_rf_inversion.py
CHANGED
|
@@ -711,7 +711,7 @@ class RFInversionFluxPipeline(
|
|
| 711 |
latents (`torch.FloatTensor`, *optional*):
|
| 712 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 713 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 714 |
-
tensor will
|
| 715 |
prompt_embeds (`torch.FloatTensor`, *optional*):
|
| 716 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 717 |
provided, text embeddings will be generated from `prompt` input argument.
|
|
|
|
| 711 |
latents (`torch.FloatTensor`, *optional*):
|
| 712 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 713 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 714 |
+
tensor will be generated by sampling using the supplied random `generator`.
|
| 715 |
prompt_embeds (`torch.FloatTensor`, *optional*):
|
| 716 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 717 |
provided, text embeddings will be generated from `prompt` input argument.
|
main/pipeline_flux_semantic_guidance.py
CHANGED
|
@@ -853,7 +853,7 @@ class FluxSemanticGuidancePipeline(
|
|
| 853 |
latents (`torch.FloatTensor`, *optional*):
|
| 854 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 855 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 856 |
-
tensor will
|
| 857 |
prompt_embeds (`torch.FloatTensor`, *optional*):
|
| 858 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 859 |
provided, text embeddings will be generated from `prompt` input argument.
|
|
|
|
| 853 |
latents (`torch.FloatTensor`, *optional*):
|
| 854 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 855 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 856 |
+
tensor will be generated by sampling using the supplied random `generator`.
|
| 857 |
prompt_embeds (`torch.FloatTensor`, *optional*):
|
| 858 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 859 |
provided, text embeddings will be generated from `prompt` input argument.
|
main/pipeline_flux_with_cfg.py
CHANGED
|
@@ -639,7 +639,7 @@ class FluxCFGPipeline(DiffusionPipeline, FluxLoraLoaderMixin, FromSingleFileMixi
|
|
| 639 |
latents (`torch.FloatTensor`, *optional*):
|
| 640 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 641 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 642 |
-
tensor will
|
| 643 |
prompt_embeds (`torch.FloatTensor`, *optional*):
|
| 644 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 645 |
provided, text embeddings will be generated from `prompt` input argument.
|
|
|
|
| 639 |
latents (`torch.FloatTensor`, *optional*):
|
| 640 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 641 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 642 |
+
tensor will be generated by sampling using the supplied random `generator`.
|
| 643 |
prompt_embeds (`torch.FloatTensor`, *optional*):
|
| 644 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 645 |
provided, text embeddings will be generated from `prompt` input argument.
|
main/pipeline_kolors_differential_img2img.py
CHANGED
|
@@ -904,7 +904,7 @@ class KolorsDifferentialImg2ImgPipeline(
|
|
| 904 |
latents (`torch.Tensor`, *optional*):
|
| 905 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 906 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 907 |
-
tensor will
|
| 908 |
prompt_embeds (`torch.Tensor`, *optional*):
|
| 909 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 910 |
provided, text embeddings will be generated from `prompt` input argument.
|
|
|
|
| 904 |
latents (`torch.Tensor`, *optional*):
|
| 905 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 906 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 907 |
+
tensor will be generated by sampling using the supplied random `generator`.
|
| 908 |
prompt_embeds (`torch.Tensor`, *optional*):
|
| 909 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 910 |
provided, text embeddings will be generated from `prompt` input argument.
|
main/pipeline_kolors_inpainting.py
CHANGED
|
@@ -1246,7 +1246,7 @@ class KolorsInpaintPipeline(
|
|
| 1246 |
latents (`torch.Tensor`, *optional*):
|
| 1247 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 1248 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 1249 |
-
tensor will
|
| 1250 |
output_type (`str`, *optional*, defaults to `"pil"`):
|
| 1251 |
The output format of the generate image. Choose between
|
| 1252 |
[PIL](https://pillow.readthedocs.io/en/stable/): `PIL.Image.Image` or `np.array`.
|
|
|
|
| 1246 |
latents (`torch.Tensor`, *optional*):
|
| 1247 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 1248 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 1249 |
+
tensor will be generated by sampling using the supplied random `generator`.
|
| 1250 |
output_type (`str`, *optional*, defaults to `"pil"`):
|
| 1251 |
The output format of the generate image. Choose between
|
| 1252 |
[PIL](https://pillow.readthedocs.io/en/stable/): `PIL.Image.Image` or `np.array`.
|
main/pipeline_prompt2prompt.py
CHANGED
|
@@ -611,7 +611,7 @@ class Prompt2PromptPipeline(
|
|
| 611 |
latents (`torch.Tensor`, *optional*):
|
| 612 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 613 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 614 |
-
tensor will
|
| 615 |
output_type (`str`, *optional*, defaults to `"pil"`):
|
| 616 |
The output format of the generate image. Choose between
|
| 617 |
[PIL](https://pillow.readthedocs.io/en/stable/): `PIL.Image.Image` or `np.array`.
|
|
|
|
| 611 |
latents (`torch.Tensor`, *optional*):
|
| 612 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 613 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 614 |
+
tensor will be generated by sampling using the supplied random `generator`.
|
| 615 |
output_type (`str`, *optional*, defaults to `"pil"`):
|
| 616 |
The output format of the generate image. Choose between
|
| 617 |
[PIL](https://pillow.readthedocs.io/en/stable/): `PIL.Image.Image` or `np.array`.
|
main/pipeline_sdxl_style_aligned.py
CHANGED
|
@@ -1480,7 +1480,7 @@ class StyleAlignedSDXLPipeline(
|
|
| 1480 |
latents (`torch.Tensor`, *optional*):
|
| 1481 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 1482 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 1483 |
-
tensor will
|
| 1484 |
prompt_embeds (`torch.Tensor`, *optional*):
|
| 1485 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 1486 |
provided, text embeddings will be generated from `prompt` input argument.
|
|
|
|
| 1480 |
latents (`torch.Tensor`, *optional*):
|
| 1481 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 1482 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 1483 |
+
tensor will be generated by sampling using the supplied random `generator`.
|
| 1484 |
prompt_embeds (`torch.Tensor`, *optional*):
|
| 1485 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 1486 |
provided, text embeddings will be generated from `prompt` input argument.
|
main/pipeline_stable_diffusion_3_differential_img2img.py
CHANGED
|
@@ -748,7 +748,7 @@ class StableDiffusion3DifferentialImg2ImgPipeline(DiffusionPipeline):
|
|
| 748 |
latents (`torch.FloatTensor`, *optional*):
|
| 749 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 750 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 751 |
-
tensor will
|
| 752 |
prompt_embeds (`torch.FloatTensor`, *optional*):
|
| 753 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 754 |
provided, text embeddings will be generated from `prompt` input argument.
|
|
|
|
| 748 |
latents (`torch.FloatTensor`, *optional*):
|
| 749 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 750 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 751 |
+
tensor will be generated by sampling using the supplied random `generator`.
|
| 752 |
prompt_embeds (`torch.FloatTensor`, *optional*):
|
| 753 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 754 |
provided, text embeddings will be generated from `prompt` input argument.
|
main/pipeline_stable_diffusion_3_instruct_pix2pix.py
CHANGED
|
@@ -945,7 +945,7 @@ class StableDiffusion3InstructPix2PixPipeline(
|
|
| 945 |
latents (`torch.FloatTensor`, *optional*):
|
| 946 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 947 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 948 |
-
tensor will
|
| 949 |
prompt_embeds (`torch.FloatTensor`, *optional*):
|
| 950 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 951 |
provided, text embeddings will be generated from `prompt` input argument.
|
|
|
|
| 945 |
latents (`torch.FloatTensor`, *optional*):
|
| 946 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 947 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 948 |
+
tensor will be generated by sampling using the supplied random `generator`.
|
| 949 |
prompt_embeds (`torch.FloatTensor`, *optional*):
|
| 950 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 951 |
provided, text embeddings will be generated from `prompt` input argument.
|
main/pipeline_stable_diffusion_xl_attentive_eraser.py
CHANGED
|
@@ -1786,7 +1786,7 @@ class StableDiffusionXL_AE_Pipeline(
|
|
| 1786 |
latents (`torch.FloatTensor`, *optional*):
|
| 1787 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 1788 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 1789 |
-
tensor will
|
| 1790 |
output_type (`str`, *optional*, defaults to `"pil"`):
|
| 1791 |
The output format of the generate image. Choose between
|
| 1792 |
[PIL](https://pillow.readthedocs.io/en/stable/): `PIL.Image.Image` or `np.array`.
|
|
|
|
| 1786 |
latents (`torch.FloatTensor`, *optional*):
|
| 1787 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 1788 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 1789 |
+
tensor will be generated by sampling using the supplied random `generator`.
|
| 1790 |
output_type (`str`, *optional*, defaults to `"pil"`):
|
| 1791 |
The output format of the generate image. Choose between
|
| 1792 |
[PIL](https://pillow.readthedocs.io/en/stable/): `PIL.Image.Image` or `np.array`.
|
main/pipeline_stable_diffusion_xl_controlnet_adapter.py
CHANGED
|
@@ -973,7 +973,7 @@ class StableDiffusionXLControlNetAdapterPipeline(
|
|
| 973 |
latents (`torch.Tensor`, *optional*):
|
| 974 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 975 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 976 |
-
tensor will
|
| 977 |
prompt_embeds (`torch.Tensor`, *optional*):
|
| 978 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 979 |
provided, text embeddings will be generated from `prompt` input argument.
|
|
|
|
| 973 |
latents (`torch.Tensor`, *optional*):
|
| 974 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 975 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 976 |
+
tensor will be generated by sampling using the supplied random `generator`.
|
| 977 |
prompt_embeds (`torch.Tensor`, *optional*):
|
| 978 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 979 |
provided, text embeddings will be generated from `prompt` input argument.
|
main/pipeline_stable_diffusion_xl_controlnet_adapter_inpaint.py
CHANGED
|
@@ -1329,7 +1329,7 @@ class StableDiffusionXLControlNetAdapterInpaintPipeline(
|
|
| 1329 |
latents (`torch.Tensor`, *optional*):
|
| 1330 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 1331 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 1332 |
-
tensor will
|
| 1333 |
prompt_embeds (`torch.Tensor`, *optional*):
|
| 1334 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 1335 |
provided, text embeddings will be generated from `prompt` input argument.
|
|
|
|
| 1329 |
latents (`torch.Tensor`, *optional*):
|
| 1330 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 1331 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 1332 |
+
tensor will be generated by sampling using the supplied random `generator`.
|
| 1333 |
prompt_embeds (`torch.Tensor`, *optional*):
|
| 1334 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 1335 |
provided, text embeddings will be generated from `prompt` input argument.
|
main/pipeline_stable_diffusion_xl_differential_img2img.py
CHANGED
|
@@ -1053,7 +1053,7 @@ class StableDiffusionXLDifferentialImg2ImgPipeline(
|
|
| 1053 |
latents (`torch.Tensor`, *optional*):
|
| 1054 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 1055 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 1056 |
-
tensor will
|
| 1057 |
prompt_embeds (`torch.Tensor`, *optional*):
|
| 1058 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 1059 |
provided, text embeddings will be generated from `prompt` input argument.
|
|
|
|
| 1053 |
latents (`torch.Tensor`, *optional*):
|
| 1054 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 1055 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 1056 |
+
tensor will be generated by sampling using the supplied random `generator`.
|
| 1057 |
prompt_embeds (`torch.Tensor`, *optional*):
|
| 1058 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 1059 |
provided, text embeddings will be generated from `prompt` input argument.
|
main/pipeline_stable_diffusion_xl_ipex.py
CHANGED
|
@@ -832,7 +832,7 @@ class StableDiffusionXLPipelineIpex(
|
|
| 832 |
latents (`torch.Tensor`, *optional*):
|
| 833 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 834 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 835 |
-
tensor will
|
| 836 |
prompt_embeds (`torch.Tensor`, *optional*):
|
| 837 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 838 |
provided, text embeddings will be generated from `prompt` input argument.
|
|
|
|
| 832 |
latents (`torch.Tensor`, *optional*):
|
| 833 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 834 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 835 |
+
tensor will be generated by sampling using the supplied random `generator`.
|
| 836 |
prompt_embeds (`torch.Tensor`, *optional*):
|
| 837 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 838 |
provided, text embeddings will be generated from `prompt` input argument.
|
main/pipeline_stg_cogvideox.py
CHANGED
|
@@ -632,7 +632,7 @@ class CogVideoXSTGPipeline(DiffusionPipeline, CogVideoXLoraLoaderMixin):
|
|
| 632 |
latents (`torch.FloatTensor`, *optional*):
|
| 633 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 634 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 635 |
-
tensor will
|
| 636 |
prompt_embeds (`torch.FloatTensor`, *optional*):
|
| 637 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 638 |
provided, text embeddings will be generated from `prompt` input argument.
|
|
|
|
| 632 |
latents (`torch.FloatTensor`, *optional*):
|
| 633 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 634 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 635 |
+
tensor will be generated by sampling using the supplied random `generator`.
|
| 636 |
prompt_embeds (`torch.FloatTensor`, *optional*):
|
| 637 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 638 |
provided, text embeddings will be generated from `prompt` input argument.
|
main/pipeline_stg_ltx.py
CHANGED
|
@@ -620,7 +620,7 @@ class LTXSTGPipeline(DiffusionPipeline, FromSingleFileMixin, LTXVideoLoraLoaderM
|
|
| 620 |
latents (`torch.Tensor`, *optional*):
|
| 621 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 622 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 623 |
-
tensor will
|
| 624 |
prompt_embeds (`torch.Tensor`, *optional*):
|
| 625 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 626 |
provided, text embeddings will be generated from `prompt` input argument.
|
|
|
|
| 620 |
latents (`torch.Tensor`, *optional*):
|
| 621 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 622 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 623 |
+
tensor will be generated by sampling using the supplied random `generator`.
|
| 624 |
prompt_embeds (`torch.Tensor`, *optional*):
|
| 625 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 626 |
provided, text embeddings will be generated from `prompt` input argument.
|
main/pipeline_stg_ltx_image2video.py
CHANGED
|
@@ -682,7 +682,7 @@ class LTXImageToVideoSTGPipeline(DiffusionPipeline, FromSingleFileMixin, LTXVide
|
|
| 682 |
latents (`torch.Tensor`, *optional*):
|
| 683 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 684 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 685 |
-
tensor will
|
| 686 |
prompt_embeds (`torch.Tensor`, *optional*):
|
| 687 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 688 |
provided, text embeddings will be generated from `prompt` input argument.
|
|
|
|
| 682 |
latents (`torch.Tensor`, *optional*):
|
| 683 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 684 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 685 |
+
tensor will be generated by sampling using the supplied random `generator`.
|
| 686 |
prompt_embeds (`torch.Tensor`, *optional*):
|
| 687 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 688 |
provided, text embeddings will be generated from `prompt` input argument.
|
main/pipeline_stg_mochi.py
CHANGED
|
@@ -603,7 +603,7 @@ class MochiSTGPipeline(DiffusionPipeline, Mochi1LoraLoaderMixin):
|
|
| 603 |
latents (`torch.Tensor`, *optional*):
|
| 604 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 605 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 606 |
-
tensor will
|
| 607 |
prompt_embeds (`torch.Tensor`, *optional*):
|
| 608 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 609 |
provided, text embeddings will be generated from `prompt` input argument.
|
|
|
|
| 603 |
latents (`torch.Tensor`, *optional*):
|
| 604 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 605 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 606 |
+
tensor will be generated by sampling using the supplied random `generator`.
|
| 607 |
prompt_embeds (`torch.Tensor`, *optional*):
|
| 608 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 609 |
provided, text embeddings will be generated from `prompt` input argument.
|
main/pipeline_zero1to3.py
CHANGED
|
@@ -657,7 +657,7 @@ class Zero1to3StableDiffusionPipeline(DiffusionPipeline, StableDiffusionMixin):
|
|
| 657 |
latents (`torch.Tensor`, *optional*):
|
| 658 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 659 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 660 |
-
tensor will
|
| 661 |
prompt_embeds (`torch.Tensor`, *optional*):
|
| 662 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 663 |
provided, text embeddings will be generated from `prompt` input argument.
|
|
|
|
| 657 |
latents (`torch.Tensor`, *optional*):
|
| 658 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 659 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 660 |
+
tensor will be generated by sampling using the supplied random `generator`.
|
| 661 |
prompt_embeds (`torch.Tensor`, *optional*):
|
| 662 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 663 |
provided, text embeddings will be generated from `prompt` input argument.
|
main/rerender_a_video.py
CHANGED
|
@@ -656,7 +656,7 @@ class RerenderAVideoPipeline(StableDiffusionControlNetImg2ImgPipeline):
|
|
| 656 |
latents (`torch.Tensor`, *optional*):
|
| 657 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 658 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 659 |
-
tensor will
|
| 660 |
prompt_embeds (`torch.Tensor`, *optional*):
|
| 661 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 662 |
provided, text embeddings will be generated from `prompt` input argument.
|
|
|
|
| 656 |
latents (`torch.Tensor`, *optional*):
|
| 657 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 658 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 659 |
+
tensor will be generated by sampling using the supplied random `generator`.
|
| 660 |
prompt_embeds (`torch.Tensor`, *optional*):
|
| 661 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 662 |
provided, text embeddings will be generated from `prompt` input argument.
|
main/run_onnx_controlnet.py
CHANGED
|
@@ -591,7 +591,7 @@ class OnnxStableDiffusionControlNetImg2ImgPipeline(DiffusionPipeline):
|
|
| 591 |
latents (`torch.Tensor`, *optional*):
|
| 592 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 593 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 594 |
-
tensor will
|
| 595 |
prompt_embeds (`torch.Tensor`, *optional*):
|
| 596 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 597 |
provided, text embeddings will be generated from `prompt` input argument.
|
|
|
|
| 591 |
latents (`torch.Tensor`, *optional*):
|
| 592 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 593 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 594 |
+
tensor will be generated by sampling using the supplied random `generator`.
|
| 595 |
prompt_embeds (`torch.Tensor`, *optional*):
|
| 596 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 597 |
provided, text embeddings will be generated from `prompt` input argument.
|
main/run_tensorrt_controlnet.py
CHANGED
|
@@ -695,7 +695,7 @@ class TensorRTStableDiffusionControlNetImg2ImgPipeline(DiffusionPipeline):
|
|
| 695 |
latents (`torch.Tensor`, *optional*):
|
| 696 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 697 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 698 |
-
tensor will
|
| 699 |
prompt_embeds (`torch.Tensor`, *optional*):
|
| 700 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 701 |
provided, text embeddings will be generated from `prompt` input argument.
|
|
|
|
| 695 |
latents (`torch.Tensor`, *optional*):
|
| 696 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 697 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 698 |
+
tensor will be generated by sampling using the supplied random `generator`.
|
| 699 |
prompt_embeds (`torch.Tensor`, *optional*):
|
| 700 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 701 |
provided, text embeddings will be generated from `prompt` input argument.
|
main/sd_text2img_k_diffusion.py
CHANGED
|
@@ -326,7 +326,7 @@ class StableDiffusionPipeline(DiffusionPipeline, StableDiffusionMixin):
|
|
| 326 |
latents (`torch.Tensor`, *optional*):
|
| 327 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 328 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 329 |
-
tensor will
|
| 330 |
output_type (`str`, *optional*, defaults to `"pil"`):
|
| 331 |
The output format of the generate image. Choose between
|
| 332 |
[PIL](https://pillow.readthedocs.io/en/stable/): `PIL.Image.Image` or `np.array`.
|
|
|
|
| 326 |
latents (`torch.Tensor`, *optional*):
|
| 327 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 328 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 329 |
+
tensor will be generated by sampling using the supplied random `generator`.
|
| 330 |
output_type (`str`, *optional*, defaults to `"pil"`):
|
| 331 |
The output format of the generate image. Choose between
|
| 332 |
[PIL](https://pillow.readthedocs.io/en/stable/): `PIL.Image.Image` or `np.array`.
|
main/seed_resize_stable_diffusion.py
CHANGED
|
@@ -122,7 +122,7 @@ class SeedResizeStableDiffusionPipeline(DiffusionPipeline, StableDiffusionMixin)
|
|
| 122 |
latents (`torch.Tensor`, *optional*):
|
| 123 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 124 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 125 |
-
tensor will
|
| 126 |
output_type (`str`, *optional*, defaults to `"pil"`):
|
| 127 |
The output format of the generate image. Choose between
|
| 128 |
[PIL](https://pillow.readthedocs.io/en/stable/): `PIL.Image.Image` or `np.array`.
|
|
|
|
| 122 |
latents (`torch.Tensor`, *optional*):
|
| 123 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 124 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 125 |
+
tensor will be generated by sampling using the supplied random `generator`.
|
| 126 |
output_type (`str`, *optional*, defaults to `"pil"`):
|
| 127 |
The output format of the generate image. Choose between
|
| 128 |
[PIL](https://pillow.readthedocs.io/en/stable/): `PIL.Image.Image` or `np.array`.
|
main/stable_diffusion_comparison.py
CHANGED
|
@@ -279,7 +279,7 @@ class StableDiffusionComparisonPipeline(DiffusionPipeline, StableDiffusionMixin)
|
|
| 279 |
latents (`torch.Tensor`, optional):
|
| 280 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 281 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 282 |
-
tensor will
|
| 283 |
output_type (`str`, optional, defaults to `"pil"`):
|
| 284 |
The output format of the generate image. Choose between
|
| 285 |
[PIL](https://pillow.readthedocs.io/en/stable/): `PIL.Image.Image` or `np.array`.
|
|
|
|
| 279 |
latents (`torch.Tensor`, optional):
|
| 280 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 281 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 282 |
+
tensor will be generated by sampling using the supplied random `generator`.
|
| 283 |
output_type (`str`, optional, defaults to `"pil"`):
|
| 284 |
The output format of the generate image. Choose between
|
| 285 |
[PIL](https://pillow.readthedocs.io/en/stable/): `PIL.Image.Image` or `np.array`.
|
main/stable_diffusion_controlnet_img2img.py
CHANGED
|
@@ -670,7 +670,7 @@ class StableDiffusionControlNetImg2ImgPipeline(DiffusionPipeline, StableDiffusio
|
|
| 670 |
latents (`torch.Tensor`, *optional*):
|
| 671 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 672 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 673 |
-
tensor will
|
| 674 |
prompt_embeds (`torch.Tensor`, *optional*):
|
| 675 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 676 |
provided, text embeddings will be generated from `prompt` input argument.
|
|
|
|
| 670 |
latents (`torch.Tensor`, *optional*):
|
| 671 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 672 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 673 |
+
tensor will be generated by sampling using the supplied random `generator`.
|
| 674 |
prompt_embeds (`torch.Tensor`, *optional*):
|
| 675 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 676 |
provided, text embeddings will be generated from `prompt` input argument.
|
main/stable_diffusion_controlnet_inpaint.py
CHANGED
|
@@ -810,7 +810,7 @@ class StableDiffusionControlNetInpaintPipeline(DiffusionPipeline, StableDiffusio
|
|
| 810 |
latents (`torch.Tensor`, *optional*):
|
| 811 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 812 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 813 |
-
tensor will
|
| 814 |
prompt_embeds (`torch.Tensor`, *optional*):
|
| 815 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 816 |
provided, text embeddings will be generated from `prompt` input argument.
|
|
|
|
| 810 |
latents (`torch.Tensor`, *optional*):
|
| 811 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 812 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 813 |
+
tensor will be generated by sampling using the supplied random `generator`.
|
| 814 |
prompt_embeds (`torch.Tensor`, *optional*):
|
| 815 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 816 |
provided, text embeddings will be generated from `prompt` input argument.
|
main/stable_diffusion_controlnet_inpaint_img2img.py
CHANGED
|
@@ -804,7 +804,7 @@ class StableDiffusionControlNetInpaintImg2ImgPipeline(DiffusionPipeline, StableD
|
|
| 804 |
latents (`torch.Tensor`, *optional*):
|
| 805 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 806 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 807 |
-
tensor will
|
| 808 |
prompt_embeds (`torch.Tensor`, *optional*):
|
| 809 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 810 |
provided, text embeddings will be generated from `prompt` input argument.
|
|
|
|
| 804 |
latents (`torch.Tensor`, *optional*):
|
| 805 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 806 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 807 |
+
tensor will be generated by sampling using the supplied random `generator`.
|
| 808 |
prompt_embeds (`torch.Tensor`, *optional*):
|
| 809 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 810 |
provided, text embeddings will be generated from `prompt` input argument.
|
main/stable_diffusion_controlnet_reference.py
CHANGED
|
@@ -179,7 +179,7 @@ class StableDiffusionControlNetReferencePipeline(StableDiffusionControlNetPipeli
|
|
| 179 |
latents (`torch.Tensor`, *optional*):
|
| 180 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 181 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 182 |
-
tensor will
|
| 183 |
prompt_embeds (`torch.Tensor`, *optional*):
|
| 184 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 185 |
provided, text embeddings will be generated from `prompt` input argument.
|
|
|
|
| 179 |
latents (`torch.Tensor`, *optional*):
|
| 180 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 181 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 182 |
+
tensor will be generated by sampling using the supplied random `generator`.
|
| 183 |
prompt_embeds (`torch.Tensor`, *optional*):
|
| 184 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 185 |
provided, text embeddings will be generated from `prompt` input argument.
|
main/stable_diffusion_ipex.py
CHANGED
|
@@ -615,7 +615,7 @@ class StableDiffusionIPEXPipeline(
|
|
| 615 |
latents (`torch.Tensor`, *optional*):
|
| 616 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 617 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 618 |
-
tensor will
|
| 619 |
prompt_embeds (`torch.Tensor`, *optional*):
|
| 620 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 621 |
provided, text embeddings will be generated from `prompt` input argument.
|
|
|
|
| 615 |
latents (`torch.Tensor`, *optional*):
|
| 616 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 617 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 618 |
+
tensor will be generated by sampling using the supplied random `generator`.
|
| 619 |
prompt_embeds (`torch.Tensor`, *optional*):
|
| 620 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 621 |
provided, text embeddings will be generated from `prompt` input argument.
|
main/stable_diffusion_reference.py
CHANGED
|
@@ -885,7 +885,7 @@ class StableDiffusionReferencePipeline(
|
|
| 885 |
latents (`torch.Tensor`, *optional*):
|
| 886 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 887 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 888 |
-
tensor will
|
| 889 |
prompt_embeds (`torch.Tensor`, *optional*):
|
| 890 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 891 |
provided, text embeddings will be generated from `prompt` input argument.
|
|
|
|
| 885 |
latents (`torch.Tensor`, *optional*):
|
| 886 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 887 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 888 |
+
tensor will be generated by sampling using the supplied random `generator`.
|
| 889 |
prompt_embeds (`torch.Tensor`, *optional*):
|
| 890 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 891 |
provided, text embeddings will be generated from `prompt` input argument.
|
main/stable_diffusion_repaint.py
CHANGED
|
@@ -678,7 +678,7 @@ class StableDiffusionRepaintPipeline(
|
|
| 678 |
latents (`torch.Tensor`, *optional*):
|
| 679 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 680 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 681 |
-
tensor will
|
| 682 |
prompt_embeds (`torch.Tensor`, *optional*):
|
| 683 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 684 |
provided, text embeddings will be generated from `prompt` input argument.
|
|
|
|
| 678 |
latents (`torch.Tensor`, *optional*):
|
| 679 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 680 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 681 |
+
tensor will be generated by sampling using the supplied random `generator`.
|
| 682 |
prompt_embeds (`torch.Tensor`, *optional*):
|
| 683 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 684 |
provided, text embeddings will be generated from `prompt` input argument.
|
main/stable_diffusion_xl_reference.py
CHANGED
|
@@ -380,7 +380,7 @@ class StableDiffusionXLReferencePipeline(StableDiffusionXLPipeline):
|
|
| 380 |
latents (`torch.Tensor`, *optional*):
|
| 381 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 382 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 383 |
-
tensor will
|
| 384 |
prompt_embeds (`torch.Tensor`, *optional*):
|
| 385 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 386 |
provided, text embeddings will be generated from `prompt` input argument.
|
|
|
|
| 380 |
latents (`torch.Tensor`, *optional*):
|
| 381 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 382 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 383 |
+
tensor will be generated by sampling using the supplied random `generator`.
|
| 384 |
prompt_embeds (`torch.Tensor`, *optional*):
|
| 385 |
Pre-generated text embeddings. Can be used to easily tweak text inputs, *e.g.* prompt weighting. If not
|
| 386 |
provided, text embeddings will be generated from `prompt` input argument.
|
main/text_inpainting.py
CHANGED
|
@@ -180,7 +180,7 @@ class TextInpainting(DiffusionPipeline, StableDiffusionMixin):
|
|
| 180 |
latents (`torch.Tensor`, *optional*):
|
| 181 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 182 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 183 |
-
tensor will
|
| 184 |
output_type (`str`, *optional*, defaults to `"pil"`):
|
| 185 |
The output format of the generate image. Choose between
|
| 186 |
[PIL](https://pillow.readthedocs.io/en/stable/): `PIL.Image.Image` or `np.array`.
|
|
|
|
| 180 |
latents (`torch.Tensor`, *optional*):
|
| 181 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 182 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 183 |
+
tensor will be generated by sampling using the supplied random `generator`.
|
| 184 |
output_type (`str`, *optional*, defaults to `"pil"`):
|
| 185 |
The output format of the generate image. Choose between
|
| 186 |
[PIL](https://pillow.readthedocs.io/en/stable/): `PIL.Image.Image` or `np.array`.
|
main/tiled_upscaling.py
CHANGED
|
@@ -231,7 +231,7 @@ class StableDiffusionTiledUpscalePipeline(StableDiffusionUpscalePipeline):
|
|
| 231 |
latents (`torch.Tensor`, *optional*):
|
| 232 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 233 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 234 |
-
tensor will
|
| 235 |
tile_size (`int`, *optional*):
|
| 236 |
The size of the tiles. Too big can result in an OOM-error.
|
| 237 |
tile_border (`int`, *optional*):
|
|
|
|
| 231 |
latents (`torch.Tensor`, *optional*):
|
| 232 |
Pre-generated noisy latents, sampled from a Gaussian distribution, to be used as inputs for image
|
| 233 |
generation. Can be used to tweak the same generation with different prompts. If not provided, a latents
|
| 234 |
+
tensor will be generated by sampling using the supplied random `generator`.
|
| 235 |
tile_size (`int`, *optional*):
|
| 236 |
The size of the tiles. Too big can result in an OOM-error.
|
| 237 |
tile_border (`int`, *optional*):
|