Spaces:
Runtime error
Runtime error
update app.py
Browse files- .gitignore +1 -2
- README.md +1 -1
- app.py +21 -26
- apps/infer.py +28 -17
- configs/econ.yaml +2 -2
- gradio_cached_examples/13/log.csv +2 -0
- gradio_cached_examples/25/log.csv +6 -0
- lib/common/libmesh/inside_mesh.py +0 -1
- lib/common/render.py +10 -3
.gitignore
CHANGED
@@ -17,5 +17,4 @@ dist
|
|
17 |
*egg-info
|
18 |
*.so
|
19 |
run.sh
|
20 |
-
*.log
|
21 |
-
gradio_cached_examples/
|
|
|
17 |
*egg-info
|
18 |
*.so
|
19 |
run.sh
|
20 |
+
*.log
|
|
README.md
CHANGED
@@ -1,6 +1,6 @@
|
|
1 |
---
|
2 |
title: Unconstrained & Detailed Clothed Human Digitization (ECON + ControlNet)
|
3 |
-
metaTitle: Avatarify from Photo
|
4 |
emoji: 🤼
|
5 |
colorFrom: green
|
6 |
colorTo: pink
|
|
|
1 |
---
|
2 |
title: Unconstrained & Detailed Clothed Human Digitization (ECON + ControlNet)
|
3 |
+
metaTitle: ECON-Avatarify from Photo
|
4 |
emoji: 🤼
|
5 |
colorFrom: green
|
6 |
colorTo: pink
|
app.py
CHANGED
@@ -6,21 +6,14 @@ import os
|
|
6 |
|
7 |
import subprocess
|
8 |
|
9 |
-
curr_dir = os.path.dirname(__file__)
|
10 |
-
|
11 |
if os.getenv('SYSTEM') == 'spaces':
|
12 |
# subprocess.run('pip install pyembree'.split())
|
13 |
subprocess.run(
|
14 |
'pip install --no-index --no-cache-dir pytorch3d -f https://dl.fbaipublicfiles.com/pytorch3d/packaging/wheels/py38_cu116_pyt1130/download.html'
|
15 |
.split()
|
16 |
)
|
17 |
-
subprocess.run(
|
18 |
-
|
19 |
-
)
|
20 |
-
subprocess.run(
|
21 |
-
f"cd {curr_dir}/lib/common/libvoxelize && python setup.py build_ext --inplace".split()
|
22 |
-
)
|
23 |
-
subprocess.run(f"cd {curr_dir}".split())
|
24 |
|
25 |
from apps.infer import generate_model, generate_video
|
26 |
|
@@ -134,6 +127,8 @@ async (image_in_img, prompt, image_file_live_opt, live_conditioning) => {
|
|
134 |
# Constants
|
135 |
low_threshold = 100
|
136 |
high_threshold = 200
|
|
|
|
|
137 |
|
138 |
# Models
|
139 |
pose_model = OpenposeDetector.from_pretrained("lllyasviel/ControlNet")
|
@@ -162,8 +157,8 @@ hint_prompts = '''
|
|
162 |
<strong>Hints</strong>: <br>
|
163 |
best quality, extremely detailed, solid color background,
|
164 |
super detail, high detail, edge lighting, soft focus,
|
165 |
-
light and dark contrast, 8k,
|
166 |
-
|
167 |
'''
|
168 |
|
169 |
|
@@ -213,8 +208,6 @@ def toggle(choice):
|
|
213 |
examples_pose = glob.glob('examples/pose/*')
|
214 |
examples_cloth = glob.glob('examples/cloth/*')
|
215 |
|
216 |
-
default_step = 50
|
217 |
-
|
218 |
with gr.Blocks() as demo:
|
219 |
gr.Markdown(description)
|
220 |
|
@@ -255,7 +248,7 @@ with gr.Blocks() as demo:
|
|
255 |
gallery_cache = gr.State()
|
256 |
inp = gr.Image(type="filepath", label="Input Image for ECON")
|
257 |
fitting_step = gr.inputs.Slider(
|
258 |
-
10, 100, step=10, label='Fitting steps', default=default_step
|
259 |
)
|
260 |
|
261 |
with gr.Row():
|
@@ -283,34 +276,34 @@ with gr.Blocks() as demo:
|
|
283 |
gr.Examples(
|
284 |
examples=list(examples_pose),
|
285 |
inputs=[inp],
|
286 |
-
cache_examples=
|
287 |
fn=generate_model,
|
288 |
outputs=out_lst,
|
289 |
-
label="Hard Pose
|
290 |
)
|
|
|
291 |
gr.Examples(
|
292 |
examples=list(examples_cloth),
|
293 |
inputs=[inp],
|
294 |
-
cache_examples=
|
295 |
fn=generate_model,
|
296 |
outputs=out_lst,
|
297 |
-
label="Loose Cloth
|
298 |
)
|
|
|
|
|
299 |
|
300 |
with gr.Column():
|
301 |
-
overlap_inp = gr.Image(type="filepath", label="Image Normal Overlap")
|
302 |
-
|
303 |
-
|
304 |
-
out_smpl = gr.Model3D(clear_color=[0.0, 0.0, 0.0, 0.0], label="SMPL-X body")
|
305 |
|
306 |
out_final_obj = gr.State()
|
307 |
vis_tensor_path = gr.State()
|
308 |
|
309 |
with gr.Row():
|
310 |
btn_video = gr.Button("Generate Video (~2min)")
|
311 |
-
|
312 |
-
out_vid = gr.Video(label="Shared on Twitter with #ECON")
|
313 |
-
|
314 |
# with gr.Row():
|
315 |
# btn_texture = gr.Button("Generate Full-texture")
|
316 |
|
@@ -345,12 +338,13 @@ with gr.Blocks() as demo:
|
|
345 |
)
|
346 |
|
347 |
btn_submit.click(fn=generate_model, inputs=[inp, fitting_step], outputs=out_lst)
|
|
|
348 |
# btn_texture.click(
|
349 |
# fn=generate_texture,
|
350 |
# inputs=[out_final_obj, prompt, seed, guidance_scale],
|
351 |
# outputs=[viewpoint_images, result_video, output_file, progress_text]
|
352 |
# )
|
353 |
-
|
354 |
demo.load(None, None, None, _js=load_js)
|
355 |
|
356 |
if __name__ == "__main__":
|
@@ -359,4 +353,5 @@ if __name__ == "__main__":
|
|
359 |
# auth=(os.environ['USER'], os.environ['PASSWORD']),
|
360 |
# auth_message="Register at icon.is.tue.mpg.de to get HuggingFace username and password.")
|
361 |
|
|
|
362 |
demo.launch(debug=True, enable_queue=True)
|
|
|
6 |
|
7 |
import subprocess
|
8 |
|
|
|
|
|
9 |
if os.getenv('SYSTEM') == 'spaces':
|
10 |
# subprocess.run('pip install pyembree'.split())
|
11 |
subprocess.run(
|
12 |
'pip install --no-index --no-cache-dir pytorch3d -f https://dl.fbaipublicfiles.com/pytorch3d/packaging/wheels/py38_cu116_pyt1130/download.html'
|
13 |
.split()
|
14 |
)
|
15 |
+
subprocess.run("python setup.py build_ext --inplace".split(), cwd="./lib/common/libmesh/")
|
16 |
+
subprocess.run("python setup.py build_ext --inplace".split(), cwd="./lib/common/libvoxelize/")
|
|
|
|
|
|
|
|
|
|
|
17 |
|
18 |
from apps.infer import generate_model, generate_video
|
19 |
|
|
|
127 |
# Constants
|
128 |
low_threshold = 100
|
129 |
high_threshold = 200
|
130 |
+
default_step = 50
|
131 |
+
cached = False
|
132 |
|
133 |
# Models
|
134 |
pose_model = OpenposeDetector.from_pretrained("lllyasviel/ControlNet")
|
|
|
157 |
<strong>Hints</strong>: <br>
|
158 |
best quality, extremely detailed, solid color background,
|
159 |
super detail, high detail, edge lighting, soft focus,
|
160 |
+
light and dark contrast, 8k, edge lighting, 3d, c4d,
|
161 |
+
blender, oc renderer, ultra high definition, 3d rendering
|
162 |
'''
|
163 |
|
164 |
|
|
|
208 |
examples_pose = glob.glob('examples/pose/*')
|
209 |
examples_cloth = glob.glob('examples/cloth/*')
|
210 |
|
|
|
|
|
211 |
with gr.Blocks() as demo:
|
212 |
gr.Markdown(description)
|
213 |
|
|
|
248 |
gallery_cache = gr.State()
|
249 |
inp = gr.Image(type="filepath", label="Input Image for ECON")
|
250 |
fitting_step = gr.inputs.Slider(
|
251 |
+
10, 100, step=10, label='Fitting steps (Slower yet Better-aligned SMPL-X)', default=default_step
|
252 |
)
|
253 |
|
254 |
with gr.Row():
|
|
|
276 |
gr.Examples(
|
277 |
examples=list(examples_pose),
|
278 |
inputs=[inp],
|
279 |
+
cache_examples=cached,
|
280 |
fn=generate_model,
|
281 |
outputs=out_lst,
|
282 |
+
label="Hard Pose Examples"
|
283 |
)
|
284 |
+
|
285 |
gr.Examples(
|
286 |
examples=list(examples_cloth),
|
287 |
inputs=[inp],
|
288 |
+
cache_examples=cached,
|
289 |
fn=generate_model,
|
290 |
outputs=out_lst,
|
291 |
+
label="Loose Cloth Examples"
|
292 |
)
|
293 |
+
|
294 |
+
out_vid = gr.Video(label="Shared on Twitter with #ECON")
|
295 |
|
296 |
with gr.Column():
|
297 |
+
overlap_inp = gr.Image(type="filepath", label="Image Normal Overlap").style(height=400)
|
298 |
+
out_final = gr.Model3D(clear_color=[0.0, 0.0, 0.0, 0.0], label="Clothed human", elem_id="avatar")
|
299 |
+
out_smpl = gr.Model3D(clear_color=[0.0, 0.0, 0.0, 0.0], label="SMPL-X body", elem_id="avatar")
|
|
|
300 |
|
301 |
out_final_obj = gr.State()
|
302 |
vis_tensor_path = gr.State()
|
303 |
|
304 |
with gr.Row():
|
305 |
btn_video = gr.Button("Generate Video (~2min)")
|
306 |
+
|
|
|
|
|
307 |
# with gr.Row():
|
308 |
# btn_texture = gr.Button("Generate Full-texture")
|
309 |
|
|
|
338 |
)
|
339 |
|
340 |
btn_submit.click(fn=generate_model, inputs=[inp, fitting_step], outputs=out_lst)
|
341 |
+
|
342 |
# btn_texture.click(
|
343 |
# fn=generate_texture,
|
344 |
# inputs=[out_final_obj, prompt, seed, guidance_scale],
|
345 |
# outputs=[viewpoint_images, result_video, output_file, progress_text]
|
346 |
# )
|
347 |
+
|
348 |
demo.load(None, None, None, _js=load_js)
|
349 |
|
350 |
if __name__ == "__main__":
|
|
|
353 |
# auth=(os.environ['USER'], os.environ['PASSWORD']),
|
354 |
# auth_message="Register at icon.is.tue.mpg.de to get HuggingFace username and password.")
|
355 |
|
356 |
+
demo.queue(concurrency_count=1)
|
357 |
demo.launch(debug=True, enable_queue=True)
|
apps/infer.py
CHANGED
@@ -28,6 +28,7 @@ import torch
|
|
28 |
import torchvision
|
29 |
import trimesh
|
30 |
from pytorch3d.ops import SubdivideMeshes
|
|
|
31 |
from termcolor import colored
|
32 |
from tqdm.auto import tqdm
|
33 |
|
@@ -47,6 +48,7 @@ from lib.net.geometry import rot6d_to_rotmat, rotation_matrix_to_angle_axis
|
|
47 |
|
48 |
torch.backends.cudnn.benchmark = True
|
49 |
|
|
|
50 |
def generate_video(vis_tensor_path):
|
51 |
|
52 |
in_tensor = torch.load(vis_tensor_path)
|
@@ -60,13 +62,14 @@ def generate_video(vis_tensor_path):
|
|
60 |
# self-rotated video
|
61 |
tmp_path = vis_tensor_path.replace("_in_tensor.pt", "_tmp.mp4")
|
62 |
out_path = vis_tensor_path.replace("_in_tensor.pt", ".mp4")
|
63 |
-
|
64 |
render.load_meshes(verts_lst, faces_lst)
|
65 |
render.get_rendered_video_multi(in_tensor, tmp_path)
|
66 |
-
|
67 |
-
os.system(f
|
68 |
-
|
69 |
-
return out_path
|
|
|
70 |
|
71 |
def generate_model(in_path, fitting_step=50):
|
72 |
|
@@ -87,7 +90,12 @@ def generate_model(in_path, fitting_step=50):
|
|
87 |
|
88 |
# load normal model
|
89 |
normal_net = Normal.load_from_checkpoint(
|
90 |
-
cfg=cfg,
|
|
|
|
|
|
|
|
|
|
|
91 |
)
|
92 |
normal_net = normal_net.to(device)
|
93 |
normal_net.netG.eval()
|
@@ -111,7 +119,12 @@ def generate_model(in_path, fitting_step=50):
|
|
111 |
if cfg.bni.use_ifnet:
|
112 |
# load IFGeo model
|
113 |
ifnet = IFGeo.load_from_checkpoint(
|
114 |
-
cfg=cfg,
|
|
|
|
|
|
|
|
|
|
|
115 |
)
|
116 |
ifnet = ifnet.to(device)
|
117 |
ifnet.netG.eval()
|
@@ -644,15 +657,13 @@ def generate_model(in_path, fitting_step=50):
|
|
644 |
overlap_path = img_overlap_path
|
645 |
vis_tensor_path = osp.join(out_dir, cfg.name, f"vid/{data['name']}_in_tensor.pt")
|
646 |
|
647 |
-
# clean all the variables
|
648 |
-
for element in dir():
|
649 |
-
|
650 |
-
|
651 |
|
652 |
-
import gc
|
653 |
-
gc.collect()
|
654 |
-
torch.cuda.empty_cache()
|
655 |
|
656 |
-
return [
|
657 |
-
smpl_glb_path, refine_glb_path, refine_obj_path, overlap_path, vis_tensor_path
|
658 |
-
]
|
|
|
28 |
import torchvision
|
29 |
import trimesh
|
30 |
from pytorch3d.ops import SubdivideMeshes
|
31 |
+
from huggingface_hub import hf_hub_download
|
32 |
from termcolor import colored
|
33 |
from tqdm.auto import tqdm
|
34 |
|
|
|
48 |
|
49 |
torch.backends.cudnn.benchmark = True
|
50 |
|
51 |
+
|
52 |
def generate_video(vis_tensor_path):
|
53 |
|
54 |
in_tensor = torch.load(vis_tensor_path)
|
|
|
62 |
# self-rotated video
|
63 |
tmp_path = vis_tensor_path.replace("_in_tensor.pt", "_tmp.mp4")
|
64 |
out_path = vis_tensor_path.replace("_in_tensor.pt", ".mp4")
|
65 |
+
|
66 |
render.load_meshes(verts_lst, faces_lst)
|
67 |
render.get_rendered_video_multi(in_tensor, tmp_path)
|
68 |
+
|
69 |
+
os.system(f"ffmpeg -y -loglevel quiet -stats -i {tmp_path} -vcodec libx264 {out_path}")
|
70 |
+
|
71 |
+
return out_path
|
72 |
+
|
73 |
|
74 |
def generate_model(in_path, fitting_step=50):
|
75 |
|
|
|
90 |
|
91 |
# load normal model
|
92 |
normal_net = Normal.load_from_checkpoint(
|
93 |
+
cfg=cfg,
|
94 |
+
checkpoint_path=hf_hub_download(
|
95 |
+
repo_id="Yuliang/ICON", use_auth_token=os.environ["ICON"], filename=cfg.normal_path
|
96 |
+
),
|
97 |
+
map_location=device,
|
98 |
+
strict=False
|
99 |
)
|
100 |
normal_net = normal_net.to(device)
|
101 |
normal_net.netG.eval()
|
|
|
119 |
if cfg.bni.use_ifnet:
|
120 |
# load IFGeo model
|
121 |
ifnet = IFGeo.load_from_checkpoint(
|
122 |
+
cfg=cfg,
|
123 |
+
checkpoint_path=hf_hub_download(
|
124 |
+
repo_id="Yuliang/ICON", use_auth_token=os.environ["ICON"], filename=cfg.ifnet_path
|
125 |
+
),
|
126 |
+
map_location=device,
|
127 |
+
strict=False
|
128 |
)
|
129 |
ifnet = ifnet.to(device)
|
130 |
ifnet.netG.eval()
|
|
|
657 |
overlap_path = img_overlap_path
|
658 |
vis_tensor_path = osp.join(out_dir, cfg.name, f"vid/{data['name']}_in_tensor.pt")
|
659 |
|
660 |
+
# # clean all the variables
|
661 |
+
# for element in dir():
|
662 |
+
# if 'path' not in element:
|
663 |
+
# del locals()[element]
|
664 |
|
665 |
+
# import gc
|
666 |
+
# gc.collect()
|
667 |
+
# torch.cuda.empty_cache()
|
668 |
|
669 |
+
return [smpl_glb_path, refine_glb_path, refine_obj_path, overlap_path, vis_tensor_path]
|
|
|
|
configs/econ.yaml
CHANGED
@@ -1,7 +1,7 @@
|
|
1 |
name: econ
|
2 |
ckpt_dir: "./data/ckpt/"
|
3 |
-
normal_path: "
|
4 |
-
ifnet_path: "
|
5 |
results_path: "./results"
|
6 |
|
7 |
net:
|
|
|
1 |
name: econ
|
2 |
ckpt_dir: "./data/ckpt/"
|
3 |
+
normal_path: "normal.ckpt"
|
4 |
+
ifnet_path: "ifnet.ckpt"
|
5 |
results_path: "./results"
|
6 |
|
7 |
net:
|
gradio_cached_examples/13/log.csv
ADDED
@@ -0,0 +1,2 @@
|
|
|
|
|
|
|
1 |
+
flag,username,timestamp
|
2 |
+
,,2023-04-15 18:15:46.412679
|
gradio_cached_examples/25/log.csv
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
flag,username,timestamp
|
2 |
+
,,2023-04-16 10:48:00.715491
|
3 |
+
,,2023-04-16 10:50:02.250539
|
4 |
+
,,2023-04-16 10:52:15.683112
|
5 |
+
,,2023-04-16 10:54:18.253116
|
6 |
+
,,2023-04-16 10:56:22.892765
|
lib/common/libmesh/inside_mesh.py
CHANGED
@@ -1,5 +1,4 @@
|
|
1 |
import numpy as np
|
2 |
-
|
3 |
from .triangle_hash import TriangleHash as _TriangleHash
|
4 |
|
5 |
|
|
|
1 |
import numpy as np
|
|
|
2 |
from .triangle_hash import TriangleHash as _TriangleHash
|
3 |
|
4 |
|
lib/common/render.py
CHANGED
@@ -38,6 +38,7 @@ from pytorch3d.renderer import (
|
|
38 |
)
|
39 |
from pytorch3d.renderer.mesh import TexturesVertex
|
40 |
from pytorch3d.structures import Meshes
|
|
|
41 |
from termcolor import colored
|
42 |
from tqdm import tqdm
|
43 |
|
@@ -305,6 +306,9 @@ class Render:
|
|
305 |
|
306 |
height, width = data["img_raw"].shape[2:]
|
307 |
|
|
|
|
|
|
|
308 |
fourcc = cv2.VideoWriter_fourcc(*"mp4v")
|
309 |
video = cv2.VideoWriter(
|
310 |
save_path,
|
@@ -351,9 +355,12 @@ class Render:
|
|
351 |
data)
|
352 |
img_cloth = blend_rgb_norm((torch.stack(mesh_renders)[num_obj:, cam_id] - 0.5) * 2.0,
|
353 |
data)
|
354 |
-
final_img = torch.cat([img_raw, img_smpl, img_cloth],
|
355 |
-
|
|
|
|
|
|
|
356 |
|
357 |
-
video.write(
|
358 |
|
359 |
video.release()
|
|
|
38 |
)
|
39 |
from pytorch3d.renderer.mesh import TexturesVertex
|
40 |
from pytorch3d.structures import Meshes
|
41 |
+
import torch.nn.functional as F
|
42 |
from termcolor import colored
|
43 |
from tqdm import tqdm
|
44 |
|
|
|
306 |
|
307 |
height, width = data["img_raw"].shape[2:]
|
308 |
|
309 |
+
width = int(width / (height / 256.0))
|
310 |
+
height = 256
|
311 |
+
|
312 |
fourcc = cv2.VideoWriter_fourcc(*"mp4v")
|
313 |
video = cv2.VideoWriter(
|
314 |
save_path,
|
|
|
355 |
data)
|
356 |
img_cloth = blend_rgb_norm((torch.stack(mesh_renders)[num_obj:, cam_id] - 0.5) * 2.0,
|
357 |
data)
|
358 |
+
final_img = torch.cat([img_raw, img_smpl, img_cloth], dim=-1).squeeze(0)
|
359 |
+
|
360 |
+
final_img_rescale = F.interpolate(
|
361 |
+
final_img, size=(height, width), mode="bilinear", align_corners=False
|
362 |
+
).squeeze(0).permute(1, 2, 0).numpy().astype(np.uint8)
|
363 |
|
364 |
+
video.write(final_img_rescale[:, :, ::-1])
|
365 |
|
366 |
video.release()
|