Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
@@ -632,14 +632,18 @@ def preprocess_img(input_image_path, raw_image_path, session_id = None):
|
|
632 |
|
633 |
def infer_example(image_path, audio_path, num_steps, raw_image_path, session_id = None, progress=gr.Progress(track_tqdm=True),):
|
634 |
|
|
|
|
|
635 |
text = ADAPTIVE_PROMPT_TEMPLATES[2]
|
636 |
|
637 |
-
result = infer(image_path, audio_path, text, num_steps,
|
638 |
|
639 |
-
|
|
|
|
|
640 |
|
641 |
@spaces.GPU(duration=get_duration)
|
642 |
-
def infer(image_path, audio_path, text, num_steps,
|
643 |
|
644 |
if image_path is None:
|
645 |
|
@@ -692,8 +696,7 @@ def infer(image_path, audio_path, text, num_steps, image_ratio, session_id = Non
|
|
692 |
|
693 |
tmp2_audio_path = os.path.join(audio_dir, f"audio_out.wav")
|
694 |
prompt_path = os.path.join(prompt_dir, f"prompt.txt")
|
695 |
-
|
696 |
-
args.image_sizes_720 = image_ratio
|
697 |
video = inferpipe(
|
698 |
prompt=text,
|
699 |
image_path=clay_image_path,
|
@@ -739,15 +742,13 @@ def orientation_changed(session_id, evt: gr.EventData):
|
|
739 |
detail = getattr(evt, "data", None) or getattr(evt, "_data", {}) or {}
|
740 |
|
741 |
if detail['value'] == "9:16":
|
742 |
-
|
743 |
elif detail['value'] == "1:1":
|
744 |
-
|
745 |
elif detail['value'] == "16:9":
|
746 |
-
|
747 |
-
|
748 |
-
print(f'{session_id} has {image_ratio} orientation')
|
749 |
|
750 |
-
|
751 |
|
752 |
def clear_raw_image():
|
753 |
return ''
|
@@ -813,7 +814,6 @@ css = """
|
|
813 |
with gr.Blocks(css=css) as demo:
|
814 |
|
815 |
session_state = gr.State()
|
816 |
-
image_ratio = gr.State([[720, 400]])
|
817 |
demo.load(start_session, outputs=[session_state])
|
818 |
|
819 |
|
@@ -911,7 +911,7 @@ with gr.Blocks(css=css) as demo:
|
|
911 |
],
|
912 |
label="Cached Examples",
|
913 |
inputs=[image_input, audio_input, num_steps, raw_img_text],
|
914 |
-
outputs=[output_video
|
915 |
fn=infer_example,
|
916 |
cache_examples=True
|
917 |
)
|
@@ -919,11 +919,11 @@ with gr.Blocks(css=css) as demo:
|
|
919 |
|
920 |
infer_btn.click(
|
921 |
fn=infer,
|
922 |
-
inputs=[image_input, audio_input, text_input, num_steps,
|
923 |
outputs=[output_video]
|
924 |
)
|
925 |
|
926 |
-
image_input.orientation(fn=orientation_changed, inputs=[session_state]
|
927 |
image_input.clear(fn=clear_raw_image, outputs=[raw_img_text])
|
928 |
image_input.upload(fn=preprocess_img, inputs=[image_input, raw_img_text, session_state], outputs=[image_input, raw_img_text])
|
929 |
image_input.change(fn=update_generate_button, inputs=[image_input, audio_input, text_input, num_steps, session_state], outputs=[time_required])
|
|
|
632 |
|
633 |
def infer_example(image_path, audio_path, num_steps, raw_image_path, session_id = None, progress=gr.Progress(track_tqdm=True),):
|
634 |
|
635 |
+
current_image_size = args.image_sizes_720
|
636 |
+
args.image_sizes_720 = [[720, 400]]
|
637 |
text = ADAPTIVE_PROMPT_TEMPLATES[2]
|
638 |
|
639 |
+
result = infer(image_path, audio_path, text, num_steps, session_id, progress)
|
640 |
|
641 |
+
args.image_sizes_720 = current_image_size
|
642 |
+
|
643 |
+
return result
|
644 |
|
645 |
@spaces.GPU(duration=get_duration)
|
646 |
+
def infer(image_path, audio_path, text, num_steps, session_id = None, progress=gr.Progress(track_tqdm=True),):
|
647 |
|
648 |
if image_path is None:
|
649 |
|
|
|
696 |
|
697 |
tmp2_audio_path = os.path.join(audio_dir, f"audio_out.wav")
|
698 |
prompt_path = os.path.join(prompt_dir, f"prompt.txt")
|
699 |
+
|
|
|
700 |
video = inferpipe(
|
701 |
prompt=text,
|
702 |
image_path=clay_image_path,
|
|
|
742 |
detail = getattr(evt, "data", None) or getattr(evt, "_data", {}) or {}
|
743 |
|
744 |
if detail['value'] == "9:16":
|
745 |
+
args.image_sizes_720 = [[720, 400]]
|
746 |
elif detail['value'] == "1:1":
|
747 |
+
args.image_sizes_720 = [[720, 720]]
|
748 |
elif detail['value'] == "16:9":
|
749 |
+
args.image_sizes_720 = [[400, 720]]
|
|
|
|
|
750 |
|
751 |
+
print(f'{session_id} has {args.image_sizes_720} orientation')
|
752 |
|
753 |
def clear_raw_image():
|
754 |
return ''
|
|
|
814 |
with gr.Blocks(css=css) as demo:
|
815 |
|
816 |
session_state = gr.State()
|
|
|
817 |
demo.load(start_session, outputs=[session_state])
|
818 |
|
819 |
|
|
|
911 |
],
|
912 |
label="Cached Examples",
|
913 |
inputs=[image_input, audio_input, num_steps, raw_img_text],
|
914 |
+
outputs=[output_video],
|
915 |
fn=infer_example,
|
916 |
cache_examples=True
|
917 |
)
|
|
|
919 |
|
920 |
infer_btn.click(
|
921 |
fn=infer,
|
922 |
+
inputs=[image_input, audio_input, text_input, num_steps, session_state],
|
923 |
outputs=[output_video]
|
924 |
)
|
925 |
|
926 |
+
image_input.orientation(fn=orientation_changed, inputs=[session_state]).then(fn=preprocess_img, inputs=[image_input, raw_img_text, session_state], outputs=[image_input, raw_img_text])
|
927 |
image_input.clear(fn=clear_raw_image, outputs=[raw_img_text])
|
928 |
image_input.upload(fn=preprocess_img, inputs=[image_input, raw_img_text, session_state], outputs=[image_input, raw_img_text])
|
929 |
image_input.change(fn=update_generate_button, inputs=[image_input, audio_input, text_input, num_steps, session_state], outputs=[time_required])
|