fffiloni commited on
Commit
7bb183a
1 Parent(s): d5175c7

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +4 -3
app.py CHANGED
@@ -255,12 +255,13 @@ with gr.Blocks(css=css) as demo:
255
  with gr.Column(elem_id="col-container"):
256
  gr.Markdown("""
257
  <h1 style="text-align: center;">ControlVideo</h1>
258
- <p> Pytorch implementation of "ControlVideo: Training-free Controllable Text-to-Video Generation" </p>
 
259
  """)
260
  with gr.Row():
261
  with gr.Column():
262
  #video_in = gr.Video(source="upload", type="filepath", visible=True)
263
- video_path = gr.Video(source="upload", type="filepath", visible=True, elem_id="video-in")
264
  prompt = gr.Textbox(label="prompt", elem_id="prompt-in")
265
  with gr.Column():
266
  video_length = gr.Slider(label="Video length", info="How many frames do you want to process ? For demo purpose, max is set to 24", minimum=1, maximum=12, step=1, value=2)
@@ -275,7 +276,7 @@ with gr.Blocks(css=css) as demo:
275
  loading_icon = gr.HTML(loading_icon_html)
276
  share_button = gr.Button("Share to community", elem_id="share-btn")
277
  gr.Examples(
278
- examples=[["Indiana Jones moonwalk in the wild jungle", "./examples/moonwalk.mp4", 'depth', 24, 424242]],
279
  fn=run_inference,
280
  inputs=[prompt,
281
  video_path,
 
255
  with gr.Column(elem_id="col-container"):
256
  gr.Markdown("""
257
  <h1 style="text-align: center;">ControlVideo</h1>
258
+ <p style="text-align: center;"> Pytorch implementation of "<a href='https://github.com/chenxwh/ControlVideo' target='_blank'>ControlVideo</a>: Training-free Controllable Text-to-Video Generation" </p>
259
+ [![Duplicate this Space](https://huggingface.co/datasets/huggingface/badges/raw/main/duplicate-this-space-sm.svg#center)](https://huggingface.co/spaces/fffiloni/ControlVideo?duplicate=true)
260
  """)
261
  with gr.Row():
262
  with gr.Column():
263
  #video_in = gr.Video(source="upload", type="filepath", visible=True)
264
+ video_path = gr.Video(label="Input video", source="upload", type="filepath", visible=True, elem_id="video-in")
265
  prompt = gr.Textbox(label="prompt", elem_id="prompt-in")
266
  with gr.Column():
267
  video_length = gr.Slider(label="Video length", info="How many frames do you want to process ? For demo purpose, max is set to 24", minimum=1, maximum=12, step=1, value=2)
 
276
  loading_icon = gr.HTML(loading_icon_html)
277
  share_button = gr.Button("Share to community", elem_id="share-btn")
278
  gr.Examples(
279
+ examples=[["Indiana Jones moonwalk in the wild jungle", "./examples/moonwalk.mp4", 'depth', 24, 192837465]],
280
  fn=run_inference,
281
  inputs=[prompt,
282
  video_path,