Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
@@ -5,7 +5,6 @@ import spaces
|
|
5 |
import torch
|
6 |
from diffusers import QwenImagePipeline
|
7 |
|
8 |
-
|
9 |
dtype = torch.bfloat16
|
10 |
device = "cuda" if torch.cuda.is_available() else "cpu"
|
11 |
|
@@ -21,8 +20,6 @@ MAX_IMAGE_SIZE = 2048
|
|
21 |
|
22 |
# pipe.flux_pipe_call_that_returns_an_iterable_of_images = flux_pipe_call_that_returns_an_iterable_of_images.__get__(pipe)
|
23 |
|
24 |
-
|
25 |
-
|
26 |
@spaces.GPU()
|
27 |
def infer(prompt, seed=42, randomize_seed=False, width=1024, height=1024, guidance_scale=4, num_inference_steps=28, lora_id=None, lora_scale=0.95, progress=gr.Progress(track_tqdm=True)):
|
28 |
if randomize_seed:
|
@@ -32,7 +29,7 @@ def infer(prompt, seed=42, randomize_seed=False, width=1024, height=1024, guidan
|
|
32 |
|
33 |
if lora_id and lora_id.strip() != "":
|
34 |
pipe.unload_lora_weights()
|
35 |
-
pipe.load_lora_weights(lora_id)
|
36 |
|
37 |
|
38 |
try:
|
|
|
5 |
import torch
|
6 |
from diffusers import QwenImagePipeline
|
7 |
|
|
|
8 |
dtype = torch.bfloat16
|
9 |
device = "cuda" if torch.cuda.is_available() else "cpu"
|
10 |
|
|
|
20 |
|
21 |
# pipe.flux_pipe_call_that_returns_an_iterable_of_images = flux_pipe_call_that_returns_an_iterable_of_images.__get__(pipe)
|
22 |
|
|
|
|
|
23 |
@spaces.GPU()
|
24 |
def infer(prompt, seed=42, randomize_seed=False, width=1024, height=1024, guidance_scale=4, num_inference_steps=28, lora_id=None, lora_scale=0.95, progress=gr.Progress(track_tqdm=True)):
|
25 |
if randomize_seed:
|
|
|
29 |
|
30 |
if lora_id and lora_id.strip() != "":
|
31 |
pipe.unload_lora_weights()
|
32 |
+
pipe.load_lora_weights(lora_id.strip())
|
33 |
|
34 |
|
35 |
try:
|