alexnasa commited on
Commit
12fea45
·
verified ·
1 Parent(s): 4b4af9d

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +15 -15
app.py CHANGED
@@ -632,14 +632,18 @@ def preprocess_img(input_image_path, raw_image_path, session_id = None):
632
 
633
  def infer_example(image_path, audio_path, num_steps, raw_image_path, session_id = None, progress=gr.Progress(track_tqdm=True),):
634
 
 
 
635
  text = ADAPTIVE_PROMPT_TEMPLATES[2]
636
 
637
- result = infer(image_path, audio_path, text, num_steps, [[720, 400]], session_id, progress)
638
 
639
- return result, image_ratio
 
 
640
 
641
  @spaces.GPU(duration=get_duration)
642
- def infer(image_path, audio_path, text, num_steps, image_ratio, session_id = None, progress=gr.Progress(track_tqdm=True),):
643
 
644
  if image_path is None:
645
 
@@ -692,8 +696,7 @@ def infer(image_path, audio_path, text, num_steps, image_ratio, session_id = Non
692
 
693
  tmp2_audio_path = os.path.join(audio_dir, f"audio_out.wav")
694
  prompt_path = os.path.join(prompt_dir, f"prompt.txt")
695
-
696
- args.image_sizes_720 = image_ratio
697
  video = inferpipe(
698
  prompt=text,
699
  image_path=clay_image_path,
@@ -739,15 +742,13 @@ def orientation_changed(session_id, evt: gr.EventData):
739
  detail = getattr(evt, "data", None) or getattr(evt, "_data", {}) or {}
740
 
741
  if detail['value'] == "9:16":
742
- image_ratio = [[720, 400]]
743
  elif detail['value'] == "1:1":
744
- image_ratio = [[720, 720]]
745
  elif detail['value'] == "16:9":
746
- image_ratio = [[400, 720]]
747
-
748
- print(f'{session_id} has {image_ratio} orientation')
749
 
750
- return image_ratio
751
 
752
  def clear_raw_image():
753
  return ''
@@ -813,7 +814,6 @@ css = """
813
  with gr.Blocks(css=css) as demo:
814
 
815
  session_state = gr.State()
816
- image_ratio = gr.State([[720, 400]])
817
  demo.load(start_session, outputs=[session_state])
818
 
819
 
@@ -911,7 +911,7 @@ with gr.Blocks(css=css) as demo:
911
  ],
912
  label="Cached Examples",
913
  inputs=[image_input, audio_input, num_steps, raw_img_text],
914
- outputs=[output_video, image_ratio],
915
  fn=infer_example,
916
  cache_examples=True
917
  )
@@ -919,11 +919,11 @@ with gr.Blocks(css=css) as demo:
919
 
920
  infer_btn.click(
921
  fn=infer,
922
- inputs=[image_input, audio_input, text_input, num_steps, image_ratio, session_state],
923
  outputs=[output_video]
924
  )
925
 
926
- image_input.orientation(fn=orientation_changed, inputs=[session_state], outputs=[image_ratio]).then(fn=preprocess_img, inputs=[image_input, raw_img_text, session_state], outputs=[image_input, raw_img_text])
927
  image_input.clear(fn=clear_raw_image, outputs=[raw_img_text])
928
  image_input.upload(fn=preprocess_img, inputs=[image_input, raw_img_text, session_state], outputs=[image_input, raw_img_text])
929
  image_input.change(fn=update_generate_button, inputs=[image_input, audio_input, text_input, num_steps, session_state], outputs=[time_required])
 
632
 
633
  def infer_example(image_path, audio_path, num_steps, raw_image_path, session_id = None, progress=gr.Progress(track_tqdm=True),):
634
 
635
+ current_image_size = args.image_sizes_720
636
+ args.image_sizes_720 = [[720, 400]]
637
  text = ADAPTIVE_PROMPT_TEMPLATES[2]
638
 
639
+ result = infer(image_path, audio_path, text, num_steps, session_id, progress)
640
 
641
+ args.image_sizes_720 = current_image_size
642
+
643
+ return result
644
 
645
  @spaces.GPU(duration=get_duration)
646
+ def infer(image_path, audio_path, text, num_steps, session_id = None, progress=gr.Progress(track_tqdm=True),):
647
 
648
  if image_path is None:
649
 
 
696
 
697
  tmp2_audio_path = os.path.join(audio_dir, f"audio_out.wav")
698
  prompt_path = os.path.join(prompt_dir, f"prompt.txt")
699
+
 
700
  video = inferpipe(
701
  prompt=text,
702
  image_path=clay_image_path,
 
742
  detail = getattr(evt, "data", None) or getattr(evt, "_data", {}) or {}
743
 
744
  if detail['value'] == "9:16":
745
+ args.image_sizes_720 = [[720, 400]]
746
  elif detail['value'] == "1:1":
747
+ args.image_sizes_720 = [[720, 720]]
748
  elif detail['value'] == "16:9":
749
+ args.image_sizes_720 = [[400, 720]]
 
 
750
 
751
+ print(f'{session_id} has {args.image_sizes_720} orientation')
752
 
753
  def clear_raw_image():
754
  return ''
 
814
  with gr.Blocks(css=css) as demo:
815
 
816
  session_state = gr.State()
 
817
  demo.load(start_session, outputs=[session_state])
818
 
819
 
 
911
  ],
912
  label="Cached Examples",
913
  inputs=[image_input, audio_input, num_steps, raw_img_text],
914
+ outputs=[output_video],
915
  fn=infer_example,
916
  cache_examples=True
917
  )
 
919
 
920
  infer_btn.click(
921
  fn=infer,
922
+ inputs=[image_input, audio_input, text_input, num_steps, session_state],
923
  outputs=[output_video]
924
  )
925
 
926
+ image_input.orientation(fn=orientation_changed, inputs=[session_state]).then(fn=preprocess_img, inputs=[image_input, raw_img_text, session_state], outputs=[image_input, raw_img_text])
927
  image_input.clear(fn=clear_raw_image, outputs=[raw_img_text])
928
  image_input.upload(fn=preprocess_img, inputs=[image_input, raw_img_text, session_state], outputs=[image_input, raw_img_text])
929
  image_input.change(fn=update_generate_button, inputs=[image_input, audio_input, text_input, num_steps, session_state], outputs=[time_required])