Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
@@ -213,11 +213,11 @@ def generate_30(
|
|
213 |
text_input_ids2 = text_inputs2.input_ids
|
214 |
|
215 |
# 2. Encode with the two text encoders
|
216 |
-
prompt_embeds_a = pipe.text_encoder(text_input_ids1, output_hidden_states=True)
|
217 |
pooled_prompt_embeds_a = prompt_embeds_a[0] # Pooled output from encoder 1
|
218 |
prompt_embeds_a = prompt_embeds_a.hidden_states[-2] # Penultimate hidden state from encoder 1
|
219 |
|
220 |
-
prompt_embeds_b = pipe.text_encoder_2(text_input_ids2, output_hidden_states=True)
|
221 |
pooled_prompt_embeds_b = prompt_embeds_b[0] # Pooled output from encoder 2
|
222 |
prompt_embeds_b = prompt_embeds_b.hidden_states[-2] # Penultimate hidden state from encoder 2
|
223 |
|
@@ -295,11 +295,11 @@ def generate_60(
|
|
295 |
text_input_ids2 = text_inputs2.input_ids
|
296 |
|
297 |
# 2. Encode with the two text encoders
|
298 |
-
prompt_embeds_a = pipe.text_encoder(text_input_ids1, output_hidden_states=True)
|
299 |
pooled_prompt_embeds_a = prompt_embeds_a[0] # Pooled output from encoder 1
|
300 |
prompt_embeds_a = prompt_embeds_a.hidden_states[-2] # Penultimate hidden state from encoder 1
|
301 |
|
302 |
-
prompt_embeds_b = pipe.text_encoder_2(text_input_ids2, output_hidden_states=True)
|
303 |
pooled_prompt_embeds_b = prompt_embeds_b[0] # Pooled output from encoder 2
|
304 |
prompt_embeds_b = prompt_embeds_b.hidden_states[-2] # Penultimate hidden state from encoder 2
|
305 |
|
@@ -377,11 +377,11 @@ def generate_90(
|
|
377 |
text_input_ids2 = text_inputs2.input_ids
|
378 |
|
379 |
# 2. Encode with the two text encoders
|
380 |
-
prompt_embeds_a = pipe.text_encoder(text_input_ids1, output_hidden_states=True)
|
381 |
pooled_prompt_embeds_a = prompt_embeds_a[0] # Pooled output from encoder 1
|
382 |
prompt_embeds_a = prompt_embeds_a.hidden_states[-2] # Penultimate hidden state from encoder 1
|
383 |
|
384 |
-
prompt_embeds_b = pipe.text_encoder_2(text_input_ids2, output_hidden_states=True)
|
385 |
pooled_prompt_embeds_b = prompt_embeds_b[0] # Pooled output from encoder 2
|
386 |
prompt_embeds_b = prompt_embeds_b.hidden_states[-2] # Penultimate hidden state from encoder 2
|
387 |
|
|
|
213 |
text_input_ids2 = text_inputs2.input_ids
|
214 |
|
215 |
# 2. Encode with the two text encoders
|
216 |
+
prompt_embeds_a = pipe.text_encoder(text_input_ids1.to(torch.device('cuda')), output_hidden_states=True)
|
217 |
pooled_prompt_embeds_a = prompt_embeds_a[0] # Pooled output from encoder 1
|
218 |
prompt_embeds_a = prompt_embeds_a.hidden_states[-2] # Penultimate hidden state from encoder 1
|
219 |
|
220 |
+
prompt_embeds_b = pipe.text_encoder_2(text_input_ids2.to(torch.device('cuda')), output_hidden_states=True)
|
221 |
pooled_prompt_embeds_b = prompt_embeds_b[0] # Pooled output from encoder 2
|
222 |
prompt_embeds_b = prompt_embeds_b.hidden_states[-2] # Penultimate hidden state from encoder 2
|
223 |
|
|
|
295 |
text_input_ids2 = text_inputs2.input_ids
|
296 |
|
297 |
# 2. Encode with the two text encoders
|
298 |
+
prompt_embeds_a = pipe.text_encoder(text_input_ids1.to(torch.device('cuda')), output_hidden_states=True)
|
299 |
pooled_prompt_embeds_a = prompt_embeds_a[0] # Pooled output from encoder 1
|
300 |
prompt_embeds_a = prompt_embeds_a.hidden_states[-2] # Penultimate hidden state from encoder 1
|
301 |
|
302 |
+
prompt_embeds_b = pipe.text_encoder_2(text_input_ids2.to(torch.device('cuda')), output_hidden_states=True)
|
303 |
pooled_prompt_embeds_b = prompt_embeds_b[0] # Pooled output from encoder 2
|
304 |
prompt_embeds_b = prompt_embeds_b.hidden_states[-2] # Penultimate hidden state from encoder 2
|
305 |
|
|
|
377 |
text_input_ids2 = text_inputs2.input_ids
|
378 |
|
379 |
# 2. Encode with the two text encoders
|
380 |
+
prompt_embeds_a = pipe.text_encoder(text_input_ids1.to(torch.device('cuda')), output_hidden_states=True)
|
381 |
pooled_prompt_embeds_a = prompt_embeds_a[0] # Pooled output from encoder 1
|
382 |
prompt_embeds_a = prompt_embeds_a.hidden_states[-2] # Penultimate hidden state from encoder 1
|
383 |
|
384 |
+
prompt_embeds_b = pipe.text_encoder_2(text_input_ids2.to(torch.device('cuda')), output_hidden_states=True)
|
385 |
pooled_prompt_embeds_b = prompt_embeds_b[0] # Pooled output from encoder 2
|
386 |
prompt_embeds_b = prompt_embeds_b.hidden_states[-2] # Penultimate hidden state from encoder 2
|
387 |
|