improve perf
Browse files
app.py
CHANGED
|
@@ -44,8 +44,10 @@ else:
|
|
| 44 |
# pipe.vae = AutoencoderTiny.from_pretrained(
|
| 45 |
# "madebyollin/taesd", torch_dtype=torch.float16, use_safetensors=True
|
| 46 |
# )
|
| 47 |
-
pipe.to(torch_device="cuda", torch_dtype=torch.float16)
|
| 48 |
pipe.set_progress_bar_config(disable=True)
|
|
|
|
|
|
|
|
|
|
| 49 |
user_queue_map = {}
|
| 50 |
|
| 51 |
|
|
@@ -55,7 +57,7 @@ def predict(input_image, prompt, guidance_scale=8.0, strength=0.5, seed=2159232)
|
|
| 55 |
num_inference_steps = 4
|
| 56 |
results = pipe(
|
| 57 |
prompt=prompt,
|
| 58 |
-
generator=generator,
|
| 59 |
image=input_image,
|
| 60 |
strength=strength,
|
| 61 |
num_inference_steps=num_inference_steps,
|
|
|
|
| 44 |
# pipe.vae = AutoencoderTiny.from_pretrained(
|
| 45 |
# "madebyollin/taesd", torch_dtype=torch.float16, use_safetensors=True
|
| 46 |
# )
|
|
|
|
| 47 |
pipe.set_progress_bar_config(disable=True)
|
| 48 |
+
pipe.to(torch_device="cuda", torch_dtype=torch.float16)
|
| 49 |
+
pipe.unet.to(memory_format=torch.channels_last)
|
| 50 |
+
pipe.unet = torch.compile(pipe.unet, mode="reduce-overhead", fullgraph=True)
|
| 51 |
user_queue_map = {}
|
| 52 |
|
| 53 |
|
|
|
|
| 57 |
num_inference_steps = 4
|
| 58 |
results = pipe(
|
| 59 |
prompt=prompt,
|
| 60 |
+
# generator=generator,
|
| 61 |
image=input_image,
|
| 62 |
strength=strength,
|
| 63 |
num_inference_steps=num_inference_steps,
|