slmnpl jbilcke-hf HF staff commited on
Commit
0f4c612
0 Parent(s):

Duplicate from jbilcke-hf/video-upscaling-server-1

Browse files

Co-authored-by: Julian Bilcke <jbilcke-hf@users.noreply.huggingface.co>

Files changed (4) hide show
  1. .gitattributes +35 -0
  2. README.md +13 -0
  3. app.py +97 -0
  4. requirements.txt +6 -0
.gitattributes ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ *.7z filter=lfs diff=lfs merge=lfs -text
2
+ *.arrow filter=lfs diff=lfs merge=lfs -text
3
+ *.bin filter=lfs diff=lfs merge=lfs -text
4
+ *.bz2 filter=lfs diff=lfs merge=lfs -text
5
+ *.ckpt filter=lfs diff=lfs merge=lfs -text
6
+ *.ftz filter=lfs diff=lfs merge=lfs -text
7
+ *.gz filter=lfs diff=lfs merge=lfs -text
8
+ *.h5 filter=lfs diff=lfs merge=lfs -text
9
+ *.joblib filter=lfs diff=lfs merge=lfs -text
10
+ *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
+ *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
+ *.model filter=lfs diff=lfs merge=lfs -text
13
+ *.msgpack filter=lfs diff=lfs merge=lfs -text
14
+ *.npy filter=lfs diff=lfs merge=lfs -text
15
+ *.npz filter=lfs diff=lfs merge=lfs -text
16
+ *.onnx filter=lfs diff=lfs merge=lfs -text
17
+ *.ot filter=lfs diff=lfs merge=lfs -text
18
+ *.parquet filter=lfs diff=lfs merge=lfs -text
19
+ *.pb filter=lfs diff=lfs merge=lfs -text
20
+ *.pickle filter=lfs diff=lfs merge=lfs -text
21
+ *.pkl filter=lfs diff=lfs merge=lfs -text
22
+ *.pt filter=lfs diff=lfs merge=lfs -text
23
+ *.pth filter=lfs diff=lfs merge=lfs -text
24
+ *.rar filter=lfs diff=lfs merge=lfs -text
25
+ *.safetensors filter=lfs diff=lfs merge=lfs -text
26
+ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
+ *.tar.* filter=lfs diff=lfs merge=lfs -text
28
+ *.tar filter=lfs diff=lfs merge=lfs -text
29
+ *.tflite filter=lfs diff=lfs merge=lfs -text
30
+ *.tgz filter=lfs diff=lfs merge=lfs -text
31
+ *.wasm filter=lfs diff=lfs merge=lfs -text
32
+ *.xz filter=lfs diff=lfs merge=lfs -text
33
+ *.zip filter=lfs diff=lfs merge=lfs -text
34
+ *.zst filter=lfs diff=lfs merge=lfs -text
35
+ *tfevents* filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ title: Zeroscope XL
3
+ emoji: 🐡
4
+ colorFrom: red
5
+ colorTo: gray
6
+ sdk: gradio
7
+ sdk_version: 3.35.2
8
+ app_file: app.py
9
+ pinned: false
10
+ duplicated_from: jbilcke-hf/video-upscaling-server-1
11
+ ---
12
+
13
+ Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference
app.py ADDED
@@ -0,0 +1,97 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import gradio as gr
2
+ import numpy as np
3
+ import os
4
+ from PIL import Image
5
+ import cv2
6
+ from moviepy.editor import VideoFileClip
7
+ import torch
8
+ from diffusers import DiffusionPipeline, DPMSolverMultistepScheduler
9
+ from diffusers.utils import export_to_video
10
+
11
+ SECRET_TOKEN = os.getenv('SECRET_TOKEN', 'default_secret')
12
+ DESCRIPTION = 'This space is an API service meant to be used by VideoChain and VideoQuest.\nWant to use this space for yourself? Please use the original code: [https://huggingface.co/spaces/fffiloni/zeroscope-XL](https://huggingface.co/spaces/fffiloni/zeroscope-XL)'
13
+
14
+ pipe_xl = DiffusionPipeline.from_pretrained("cerspense/zeroscope_v2_XL", torch_dtype=torch.float16, revision="refs/pr/17")
15
+ pipe_xl.vae.enable_slicing()
16
+ pipe_xl.scheduler = DPMSolverMultistepScheduler.from_config(pipe_xl.scheduler.config)
17
+ pipe_xl.enable_model_cpu_offload()
18
+ pipe_xl.to("cuda")
19
+
20
+ def convert_mp4_to_frames(video_path):
21
+ # Read the video file
22
+ video = cv2.VideoCapture(video_path)
23
+
24
+ # Get the frames per second (fps) of the video
25
+ fps = video.get(cv2.CAP_PROP_FPS)
26
+
27
+ # Calculate the number of frames to extract
28
+ # Note: we cannot go beyond 3 seconds on the large A10G
29
+ # num_frames = int(fps * min(duration, 3))
30
+
31
+ # let's use a fixed max for now, no need to have a duration field
32
+ num_frames = int(fps * 3)
33
+
34
+ frames = []
35
+ frame_count = 0
36
+
37
+ # Iterate through each frame
38
+ while True:
39
+ # Read a frame
40
+ ret, frame = video.read()
41
+
42
+ # If the frame was not successfully read or we have reached the desired duration, break the loop
43
+ if not ret or frame_count == num_frames:
44
+ break
45
+
46
+ # Convert BGR to RGB
47
+ frame = cv2.cvtColor(frame, cv2.COLOR_BGR2RGB)
48
+
49
+ # Append the frame to the list of frames
50
+ frames.append(frame)
51
+
52
+ frame_count += 1
53
+
54
+ # Release the video object
55
+ video.release()
56
+
57
+ # Convert the list of frames to a numpy array
58
+ frames = np.array(frames)
59
+
60
+ return frames
61
+
62
+ def infer(prompt, video_in, denoise_strength, secret_token):
63
+ if secret_token != SECRET_TOKEN:
64
+ raise gr.Error(
65
+ f'Invalid secret token. Please fork the original space if you want to use it for yourself.')
66
+
67
+ negative_prompt = "text, watermark, copyright, blurry, cropped, noisy, pixelated, nsfw"
68
+
69
+
70
+ video = convert_mp4_to_frames(video_in)
71
+ video_resized = [Image.fromarray(frame).resize((1024, 576)) for frame in video]
72
+ video_frames = pipe_xl(prompt, negative_prompt=negative_prompt, video=video_resized, strength=denoise_strength).frames
73
+ video_path = export_to_video(video_frames, output_video_path="xl_result.mp4")
74
+
75
+ return "xl_result.mp4", gr.Group.update(visible=True)
76
+
77
+
78
+ with gr.Blocks() as demo:
79
+ gr.Markdown(DESCRIPTION)
80
+
81
+ with gr.Column():
82
+ secret_token = gr.Text(label='Secret Token', max_lines=1)
83
+ video_in = gr.Video(type="numpy", source="upload")
84
+ prompt_in = gr.Textbox(label="Prompt", elem_id="prompt-in")
85
+ denoise_strength = gr.Slider(label="Denoise strength", minimum=0.6, maximum=0.9, step=0.01, value=0.66)
86
+ #inference_steps = gr.Slider(label="Inference Steps", minimum=7, maximum=100, step=1, value=40, interactive=False)
87
+ submit_btn = gr.Button("Submit")
88
+ video_result = gr.Video(label="Video Output", elem_id="video-output")
89
+
90
+ submit_btn.click(fn=infer,
91
+ inputs=[prompt_in, video_in, denoise_strength, secret_token],
92
+ outputs=[video_result],
93
+ api_name="zero_xl"
94
+ )
95
+
96
+ demo.queue(max_size=6).launch()
97
+
requirements.txt ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ git+https://github.com/huggingface/diffusers
2
+ transformers
3
+ accelerate
4
+ torch
5
+ opencv-python
6
+ moviepy