Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
@@ -269,6 +269,9 @@ def generate_30(
|
|
269 |
# 3. Concatenate the text_encoder_2 embeddings
|
270 |
prompt_embeds2 = torch.cat([prompt_embeds_a2, prompt_embeds_b2])
|
271 |
print('catted shape2: ', prompt_embeds2.shape)
|
|
|
|
|
|
|
272 |
pooled_prompt_embeds2 = torch.cat([pooled_prompt_embeds_a2, pooled_prompt_embeds_b2])
|
273 |
print('catted pooled shape 2: ', pooled_prompt_embeds2.shape)
|
274 |
pooled_prompt_embeds2 = torch.mean(pooled_prompt_embeds2,dim=0,keepdim=True)
|
@@ -276,6 +279,9 @@ def generate_30(
|
|
276 |
pooled_prompt_embeds = torch.cat([pooled_prompt_embeds, pooled_prompt_embeds2],dim=1)
|
277 |
print('catted combined meaned pooled shape: ', pooled_prompt_embeds.shape)
|
278 |
|
|
|
|
|
|
|
279 |
options = {
|
280 |
#"prompt": prompt,
|
281 |
"prompt_embeds": prompt_embeds,
|
@@ -398,6 +404,9 @@ def generate_60(
|
|
398 |
# 3. Concatenate the text_encoder_2 embeddings
|
399 |
prompt_embeds2 = torch.cat([prompt_embeds_a2, prompt_embeds_b2])
|
400 |
print('catted shape2: ', prompt_embeds2.shape)
|
|
|
|
|
|
|
401 |
pooled_prompt_embeds2 = torch.cat([pooled_prompt_embeds_a2, pooled_prompt_embeds_b2])
|
402 |
print('catted pooled shape 2: ', pooled_prompt_embeds2.shape)
|
403 |
pooled_prompt_embeds2 = torch.mean(pooled_prompt_embeds2,dim=0,keepdim=True)
|
@@ -405,6 +414,9 @@ def generate_60(
|
|
405 |
pooled_prompt_embeds = torch.cat([pooled_prompt_embeds, pooled_prompt_embeds2],dim=1)
|
406 |
print('catted combined meaned pooled shape: ', pooled_prompt_embeds.shape)
|
407 |
|
|
|
|
|
|
|
408 |
options = {
|
409 |
#"prompt": prompt,
|
410 |
"prompt_embeds": prompt_embeds,
|
@@ -527,6 +539,9 @@ def generate_90(
|
|
527 |
# 3. Concatenate the text_encoder_2 embeddings
|
528 |
prompt_embeds2 = torch.cat([prompt_embeds_a2, prompt_embeds_b2])
|
529 |
print('catted shape2: ', prompt_embeds2.shape)
|
|
|
|
|
|
|
530 |
pooled_prompt_embeds2 = torch.cat([pooled_prompt_embeds_a2, pooled_prompt_embeds_b2])
|
531 |
print('catted pooled shape 2: ', pooled_prompt_embeds2.shape)
|
532 |
pooled_prompt_embeds2 = torch.mean(pooled_prompt_embeds2,dim=0,keepdim=True)
|
@@ -534,6 +549,9 @@ def generate_90(
|
|
534 |
pooled_prompt_embeds = torch.cat([pooled_prompt_embeds, pooled_prompt_embeds2],dim=1)
|
535 |
print('catted combined meaned pooled shape: ', pooled_prompt_embeds.shape)
|
536 |
|
|
|
|
|
|
|
537 |
options = {
|
538 |
#"prompt": prompt,
|
539 |
"prompt_embeds": prompt_embeds,
|
|
|
269 |
# 3. Concatenate the text_encoder_2 embeddings
|
270 |
prompt_embeds2 = torch.cat([prompt_embeds_a2, prompt_embeds_b2])
|
271 |
print('catted shape2: ', prompt_embeds2.shape)
|
272 |
+
prompt_embeds2 = torch.mean(prompt_embeds2,dim=0,keepdim=True)
|
273 |
+
print('averaged shape2: ', prompt_embeds.shape)
|
274 |
+
|
275 |
pooled_prompt_embeds2 = torch.cat([pooled_prompt_embeds_a2, pooled_prompt_embeds_b2])
|
276 |
print('catted pooled shape 2: ', pooled_prompt_embeds2.shape)
|
277 |
pooled_prompt_embeds2 = torch.mean(pooled_prompt_embeds2,dim=0,keepdim=True)
|
|
|
279 |
pooled_prompt_embeds = torch.cat([pooled_prompt_embeds, pooled_prompt_embeds2],dim=1)
|
280 |
print('catted combined meaned pooled shape: ', pooled_prompt_embeds.shape)
|
281 |
|
282 |
+
prompt_embeds = torch.cat([prompt_embeds, prompt_embeds2],dim=-1)
|
283 |
+
print('catted combined meaned shape: ', prompt_embeds.shape)
|
284 |
+
|
285 |
options = {
|
286 |
#"prompt": prompt,
|
287 |
"prompt_embeds": prompt_embeds,
|
|
|
404 |
# 3. Concatenate the text_encoder_2 embeddings
|
405 |
prompt_embeds2 = torch.cat([prompt_embeds_a2, prompt_embeds_b2])
|
406 |
print('catted shape2: ', prompt_embeds2.shape)
|
407 |
+
prompt_embeds2 = torch.mean(prompt_embeds2,dim=0,keepdim=True)
|
408 |
+
print('averaged shape2: ', prompt_embeds.shape)
|
409 |
+
|
410 |
pooled_prompt_embeds2 = torch.cat([pooled_prompt_embeds_a2, pooled_prompt_embeds_b2])
|
411 |
print('catted pooled shape 2: ', pooled_prompt_embeds2.shape)
|
412 |
pooled_prompt_embeds2 = torch.mean(pooled_prompt_embeds2,dim=0,keepdim=True)
|
|
|
414 |
pooled_prompt_embeds = torch.cat([pooled_prompt_embeds, pooled_prompt_embeds2],dim=1)
|
415 |
print('catted combined meaned pooled shape: ', pooled_prompt_embeds.shape)
|
416 |
|
417 |
+
prompt_embeds = torch.cat([prompt_embeds, prompt_embeds2],dim=-1)
|
418 |
+
print('catted combined meaned shape: ', prompt_embeds.shape)
|
419 |
+
|
420 |
options = {
|
421 |
#"prompt": prompt,
|
422 |
"prompt_embeds": prompt_embeds,
|
|
|
539 |
# 3. Concatenate the text_encoder_2 embeddings
|
540 |
prompt_embeds2 = torch.cat([prompt_embeds_a2, prompt_embeds_b2])
|
541 |
print('catted shape2: ', prompt_embeds2.shape)
|
542 |
+
prompt_embeds2 = torch.mean(prompt_embeds2,dim=0,keepdim=True)
|
543 |
+
print('averaged shape2: ', prompt_embeds.shape)
|
544 |
+
|
545 |
pooled_prompt_embeds2 = torch.cat([pooled_prompt_embeds_a2, pooled_prompt_embeds_b2])
|
546 |
print('catted pooled shape 2: ', pooled_prompt_embeds2.shape)
|
547 |
pooled_prompt_embeds2 = torch.mean(pooled_prompt_embeds2,dim=0,keepdim=True)
|
|
|
549 |
pooled_prompt_embeds = torch.cat([pooled_prompt_embeds, pooled_prompt_embeds2],dim=1)
|
550 |
print('catted combined meaned pooled shape: ', pooled_prompt_embeds.shape)
|
551 |
|
552 |
+
prompt_embeds = torch.cat([prompt_embeds, prompt_embeds2],dim=-1)
|
553 |
+
print('catted combined meaned shape: ', prompt_embeds.shape)
|
554 |
+
|
555 |
options = {
|
556 |
#"prompt": prompt,
|
557 |
"prompt_embeds": prompt_embeds,
|