Shariq F. Bhat commited on
Commit
99b3515
1 Parent(s): 3622d44

initial release

Browse files
.gitattributes CHANGED
@@ -32,3 +32,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
32
  *.zip filter=lfs diff=lfs merge=lfs -text
33
  *.zst filter=lfs diff=lfs merge=lfs -text
34
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
32
  *.zip filter=lfs diff=lfs merge=lfs -text
33
  *.zst filter=lfs diff=lfs merge=lfs -text
34
  *tfevents* filter=lfs diff=lfs merge=lfs -text
35
+ examples/person-leaves.png filter=lfs diff=lfs merge=lfs -text
.gitignore ADDED
@@ -0,0 +1,2 @@
 
 
1
+ __pycache__/
2
+ **.pyc
app.py ADDED
@@ -0,0 +1,42 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import gradio as gr
2
+ import torch
3
+
4
+ from gradio_depth_pred import create_demo as create_depth_pred_demo
5
+ from gradio_im_to_3d import create_demo as create_im_to_3d_demo
6
+ from gradio_pano_to_3d import create_demo as create_pano_to_3d_demo
7
+
8
+
9
+ css = """
10
+ #img-display-container {
11
+ max-height: 50vh;
12
+ }
13
+ #img-display-input {
14
+ max-height: 40vh;
15
+ }
16
+ #img-display-output {
17
+ max-height: 40vh;
18
+ }
19
+
20
+ """
21
+ DEVICE = 'cuda' if torch.cuda.is_available() else 'cpu'
22
+ model = torch.hub.load('isl-org/ZoeDepth', "ZoeD_N", pretrained=True).to(DEVICE).eval()
23
+
24
+ title = "# ZoeDepth"
25
+ description = """Official demo for **ZoeDepth: Zero-shot Transfer by Combining Relative and Metric Depth**.
26
+
27
+ ZoeDepth is a deep learning model for metric depth estimation from a single image.
28
+
29
+ Please refer to our [paper](https://arxiv.org/abs/2302.12288) or [github](https://github.com/isl-org/ZoeDepth) for more details."""
30
+
31
+ with gr.Blocks(css=css) as demo:
32
+ gr.Markdown(title)
33
+ gr.Markdown(description)
34
+ with gr.Tab("Depth Prediction"):
35
+ create_depth_pred_demo(model)
36
+ with gr.Tab("Image to 3D"):
37
+ create_im_to_3d_demo(model)
38
+ with gr.Tab("360 Panorama to 3D"):
39
+ create_pano_to_3d_demo(model)
40
+
41
+ if __name__ == '__main__':
42
+ demo.queue().launch()
examples/aerial_beach.jpeg ADDED
examples/ancient-carved.jpeg ADDED
examples/living-room.jpeg ADDED
examples/mountains.jpeg ADDED
examples/pano_1.jpeg ADDED
examples/pano_2.jpeg ADDED
examples/pano_3.jpeg ADDED
examples/person-leaves.png ADDED

Git LFS Details

  • SHA256: 00791e189fd3df0b963288332a944c420a606891aba7cc2d6356bdd7378fbb7b
  • Pointer size: 132 Bytes
  • Size of remote file: 1.18 MB
examples/person_1.jpeg ADDED
examples/person_2.jpeg ADDED
geometry.py ADDED
@@ -0,0 +1,72 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import numpy as np
2
+
3
+ def get_intrinsics(H,W):
4
+ """
5
+ Intrinsics for a pinhole camera model.
6
+ Assume fov of 55 degrees and central principal point.
7
+ """
8
+ f = 0.5 * W / np.tan(0.5 * 55 * np.pi / 180.0)
9
+ cx = 0.5 * W
10
+ cy = 0.5 * H
11
+ return np.array([[f, 0, cx],
12
+ [0, f, cy],
13
+ [0, 0, 1]])
14
+
15
+ def depth_to_points(depth, R=None, t=None):
16
+
17
+ K = get_intrinsics(depth.shape[1], depth.shape[2])
18
+ Kinv = np.linalg.inv(K)
19
+ if R is None:
20
+ R = np.eye(3)
21
+ if t is None:
22
+ t = np.zeros(3)
23
+
24
+ # M converts from your coordinate to PyTorch3D's coordinate system
25
+ M = np.eye(3)
26
+ M[0, 0] = -1.0
27
+ M[1, 1] = -1.0
28
+
29
+ height, width = depth.shape[1:3]
30
+
31
+ x = np.arange(width)
32
+ y = np.arange(height)
33
+ coord = np.stack(np.meshgrid(x, y), -1)
34
+ coord = np.concatenate((coord, np.ones_like(coord)[:, :, [0]]), -1) # z=1
35
+ coord = coord.astype(np.float32)
36
+ # coord = torch.as_tensor(coord, dtype=torch.float32, device=device)
37
+ coord = coord[None] # bs, h, w, 3
38
+
39
+ D = depth[:, :, :, None, None]
40
+ # print(D.shape, Kinv[None, None, None, ...].shape, coord[:, :, :, :, None].shape )
41
+ pts3D_1 = D * Kinv[None, None, None, ...] @ coord[:, :, :, :, None]
42
+ # pts3D_1 live in your coordinate system. Convert them to Py3D's
43
+ pts3D_1 = M[None, None, None, ...] @ pts3D_1
44
+ # from reference to targe tviewpoint
45
+ pts3D_2 = R[None, None, None, ...] @ pts3D_1 + t[None, None, None, :, None]
46
+ # pts3D_2 = pts3D_1
47
+ # depth_2 = pts3D_2[:, :, :, 2, :] # b,1,h,w
48
+ return pts3D_2[:, :, :, :3, 0][0]
49
+
50
+
51
+ def create_triangles(h, w, mask=None):
52
+ """Creates mesh triangle indices from a given pixel grid size.
53
+ This function is not and need not be differentiable as triangle indices are
54
+ fixed.
55
+ Args:
56
+ h: (int) denoting the height of the image.
57
+ w: (int) denoting the width of the image.
58
+ Returns:
59
+ triangles: 2D numpy array of indices (int) with shape (2(W-1)(H-1) x 3)
60
+ """
61
+ x, y = np.meshgrid(range(w - 1), range(h - 1))
62
+ tl = y * w + x
63
+ tr = y * w + x + 1
64
+ bl = (y + 1) * w + x
65
+ br = (y + 1) * w + x + 1
66
+ triangles = np.array([tl, bl, tr, br, tr, bl])
67
+ triangles = np.transpose(triangles, (1, 2, 0)).reshape(
68
+ ((w - 1) * (h - 1) * 2, 3))
69
+ if mask is not None:
70
+ mask = mask.reshape(-1)
71
+ triangles = triangles[mask[triangles].all(1)]
72
+ return triangles
gradio_depth_pred.py ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import gradio as gr
2
+ from utils import colorize
3
+ from PIL import Image
4
+ import tempfile
5
+
6
+ def predict_depth(model, image):
7
+ depth = model.infer_pil(image)
8
+ return depth
9
+
10
+ def create_demo(model):
11
+ gr.Markdown("### Depth Prediction demo")
12
+ with gr.Row():
13
+ input_image = gr.Image(label="Input Image", type='pil', elem_id='img-display-input').style(height="auto")
14
+ depth_image = gr.Image(label="Depth Map", elem_id='img-display-output')
15
+ raw_file = gr.File(label="16-bit raw depth, multiplier:256")
16
+ submit = gr.Button("Submit")
17
+
18
+ def on_submit(image):
19
+ depth = predict_depth(model, image)
20
+ colored_depth = colorize(depth, cmap='gray_r')
21
+ tmp = tempfile.NamedTemporaryFile(suffix='.png', delete=False)
22
+ raw_depth = Image.fromarray((depth*256).astype('uint16'))
23
+ raw_depth.save(tmp.name)
24
+ return [colored_depth, tmp.name]
25
+
26
+ submit.click(on_submit, inputs=[input_image], outputs=[depth_image, raw_file])
27
+ examples = gr.Examples(examples=["examples/person_1.jpeg", "examples/person_2.jpeg", "examples/person-leaves.png", "examples/living-room.jpeg"],
28
+ inputs=[input_image])
gradio_im_to_3d.py ADDED
@@ -0,0 +1,69 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import gradio as gr
2
+ import numpy as np
3
+ import trimesh
4
+ from geometry import depth_to_points, create_triangles
5
+ from functools import partial
6
+ import tempfile
7
+
8
+
9
+ def depth_edges_mask(depth):
10
+ """Returns a mask of edges in the depth map.
11
+ Args:
12
+ depth: 2D numpy array of shape (H, W) with dtype float32.
13
+ Returns:
14
+ mask: 2D numpy array of shape (H, W) with dtype bool.
15
+ """
16
+ # Compute the x and y gradients of the depth map.
17
+ depth_dx, depth_dy = np.gradient(depth)
18
+ # Compute the gradient magnitude.
19
+ depth_grad = np.sqrt(depth_dx ** 2 + depth_dy ** 2)
20
+ # Compute the edge mask.
21
+ mask = depth_grad > 0.05
22
+ return mask
23
+
24
+
25
+ def predict_depth(model, image):
26
+ depth = model.infer_pil(image)
27
+ return depth
28
+
29
+ def get_mesh(model, image, keep_edges=False):
30
+ image.thumbnail((1024,1024)) # limit the size of the input image
31
+ depth = predict_depth(model, image)
32
+ pts3d = depth_to_points(depth[None])
33
+ pts3d = pts3d.reshape(-1, 3)
34
+
35
+ # Create a trimesh mesh from the points
36
+ # Each pixel is connected to its 4 neighbors
37
+ # colors are the RGB values of the image
38
+
39
+ verts = pts3d.reshape(-1, 3)
40
+ image = np.array(image)
41
+ if keep_edges:
42
+ triangles = create_triangles(image.shape[0], image.shape[1])
43
+ else:
44
+ triangles = create_triangles(image.shape[0], image.shape[1], mask=~depth_edges_mask(depth))
45
+ colors = image.reshape(-1, 3)
46
+ mesh = trimesh.Trimesh(vertices=verts, faces=triangles, vertex_colors=colors)
47
+
48
+ # Save as glb
49
+ glb_file = tempfile.NamedTemporaryFile(suffix='.glb', delete=False)
50
+ glb_path = glb_file.name
51
+ mesh.export(glb_path)
52
+ return glb_path
53
+
54
+ def create_demo(model):
55
+
56
+ gr.Markdown("### Image to 3D mesh")
57
+ gr.Markdown("Convert a single 2D image to a 3D mesh")
58
+
59
+ with gr.Row():
60
+ image = gr.Image(label="Input Image", type='pil')
61
+ result = gr.Model3D(label="3d mesh reconstruction", clear_color=[
62
+ 1.0, 1.0, 1.0, 1.0])
63
+
64
+ checkbox = gr.Checkbox(label="Keep occlusion edges", value=False)
65
+ submit = gr.Button("Submit")
66
+ submit.click(partial(get_mesh, model), inputs=[image, checkbox], outputs=[result])
67
+ examples = gr.Examples(examples=["examples/aerial_beach.jpeg", "examples/mountains.jpeg", "examples/person_1.jpeg", "examples/ancient-carved.jpeg"],
68
+ inputs=[image])
69
+
gradio_pano_to_3d.py ADDED
@@ -0,0 +1,96 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import gradio as gr
2
+ import numpy as np
3
+ import trimesh
4
+ from geometry import create_triangles
5
+ from functools import partial
6
+ import tempfile
7
+
8
+ def depth_edges_mask(depth):
9
+ """Returns a mask of edges in the depth map.
10
+ Args:
11
+ depth: 2D numpy array of shape (H, W) with dtype float32.
12
+ Returns:
13
+ mask: 2D numpy array of shape (H, W) with dtype bool.
14
+ """
15
+ # Compute the x and y gradients of the depth map.
16
+ depth_dx, depth_dy = np.gradient(depth)
17
+ # Compute the gradient magnitude.
18
+ depth_grad = np.sqrt(depth_dx ** 2 + depth_dy ** 2)
19
+ # Compute the edge mask.
20
+ mask = depth_grad > 0.05
21
+ return mask
22
+
23
+
24
+ def pano_depth_to_world_points(depth):
25
+ """
26
+ 360 depth to world points
27
+ given 2D depth is an equirectangular projection of a spherical image
28
+ Treat depth as radius
29
+
30
+ longitude : -pi to pi
31
+ latitude : -pi/2 to pi/2
32
+ """
33
+
34
+ # Convert depth to radius
35
+ radius = depth.flatten()
36
+
37
+ lon = np.linspace(-np.pi, np.pi, depth.shape[1])
38
+ lat = np.linspace(-np.pi/2, np.pi/2, depth.shape[0])
39
+
40
+ lon, lat = np.meshgrid(lon, lat)
41
+ lon = lon.flatten()
42
+ lat = lat.flatten()
43
+
44
+ # Convert to cartesian coordinates
45
+ x = radius * np.cos(lat) * np.cos(lon)
46
+ y = radius * np.cos(lat) * np.sin(lon)
47
+ z = radius * np.sin(lat)
48
+
49
+ pts3d = np.stack([x, y, z], axis=1)
50
+
51
+ return pts3d
52
+
53
+
54
+ def predict_depth(model, image):
55
+ depth = model.infer_pil(image)
56
+ return depth
57
+
58
+ def get_mesh(model, image, keep_edges=False):
59
+ image.thumbnail((1024,1024)) # limit the size of the image
60
+ depth = predict_depth(model, image)
61
+ pts3d = pano_depth_to_world_points(depth)
62
+
63
+ # Create a trimesh mesh from the points
64
+ # Each pixel is connected to its 4 neighbors
65
+ # colors are the RGB values of the image
66
+
67
+ verts = pts3d.reshape(-1, 3)
68
+ image = np.array(image)
69
+ if keep_edges:
70
+ triangles = create_triangles(image.shape[0], image.shape[1])
71
+ else:
72
+ triangles = create_triangles(image.shape[0], image.shape[1], mask=~depth_edges_mask(depth))
73
+ colors = image.reshape(-1, 3)
74
+ mesh = trimesh.Trimesh(vertices=verts, faces=triangles, vertex_colors=colors)
75
+
76
+ # Save as glb
77
+ glb_file = tempfile.NamedTemporaryFile(suffix='.glb', delete=False)
78
+ glb_path = glb_file.name
79
+ mesh.export(glb_path)
80
+ return glb_path
81
+
82
+ def create_demo(model):
83
+ gr.Markdown("### Panorama to 3D mesh")
84
+ gr.Markdown("Convert a 360 spherical panorama to a 3D mesh")
85
+ gr.Markdown("ZoeDepth was not trained on panoramic images. It doesn't know anything about panoramas or spherical projection. Here, we just treat the estimated depth as radius and some projection errors are expected. Nonetheless, ZoeDepth still works surprisingly well on 360 reconstruction.")
86
+
87
+ with gr.Row():
88
+ input_image = gr.Image(label="Input Image", type='pil')
89
+ result = gr.Model3D(label="3d mesh reconstruction", clear_color=[
90
+ 1.0, 1.0, 1.0, 1.0])
91
+
92
+ checkbox = gr.Checkbox(label="Keep occlusion edges", value=True)
93
+ submit = gr.Button("Submit")
94
+ submit.click(partial(get_mesh, model), inputs=[input_image, checkbox], outputs=[result])
95
+ examples = gr.Examples(examples=["examples/pano_1.jpeg", "examples/pano_2.jpeg", "examples/pano_3.jpeg"],
96
+ inputs=[input_image])
requirements.txt ADDED
@@ -0,0 +1,4 @@
 
 
 
 
1
+ timm==0.6.11
2
+ torch==1.10.1
3
+ torchvision==0.11.2
4
+ trimesh==3.9.42
utils.py ADDED
@@ -0,0 +1,86 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # MIT License
2
+
3
+ # Copyright (c) 2022 Intelligent Systems Lab Org
4
+
5
+ # Permission is hereby granted, free of charge, to any person obtaining a copy
6
+ # of this software and associated documentation files (the "Software"), to deal
7
+ # in the Software without restriction, including without limitation the rights
8
+ # to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
9
+ # copies of the Software, and to permit persons to whom the Software is
10
+ # furnished to do so, subject to the following conditions:
11
+
12
+ # The above copyright notice and this permission notice shall be included in all
13
+ # copies or substantial portions of the Software.
14
+
15
+ # THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
16
+ # IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
17
+ # FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
18
+ # AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
19
+ # LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
20
+ # OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
21
+ # SOFTWARE.
22
+
23
+ # File author: Shariq Farooq Bhat
24
+
25
+ import matplotlib
26
+ import matplotlib.cm
27
+ import numpy as np
28
+ import torch
29
+
30
+ def colorize(value, vmin=None, vmax=None, cmap='magma_r', invalid_val=-99, invalid_mask=None, background_color=(128, 128, 128, 255), gamma_corrected=False, value_transform=None):
31
+ """Converts a depth map to a color image.
32
+
33
+ Args:
34
+ value (torch.Tensor, numpy.ndarry): Input depth map. Shape: (H, W) or (1, H, W) or (1, 1, H, W). All singular dimensions are squeezed
35
+ vmin (float, optional): vmin-valued entries are mapped to start color of cmap. If None, value.min() is used. Defaults to None.
36
+ vmax (float, optional): vmax-valued entries are mapped to end color of cmap. If None, value.max() is used. Defaults to None.
37
+ cmap (str, optional): matplotlib colormap to use. Defaults to 'magma_r'.
38
+ invalid_val (int, optional): Specifies value of invalid pixels that should be colored as 'background_color'. Defaults to -99.
39
+ invalid_mask (numpy.ndarray, optional): Boolean mask for invalid regions. Defaults to None.
40
+ background_color (tuple[int], optional): 4-tuple RGB color to give to invalid pixels. Defaults to (128, 128, 128, 255).
41
+ gamma_corrected (bool, optional): Apply gamma correction to colored image. Defaults to False.
42
+ value_transform (Callable, optional): Apply transform function to valid pixels before coloring. Defaults to None.
43
+
44
+ Returns:
45
+ numpy.ndarray, dtype - uint8: Colored depth map. Shape: (H, W, 4)
46
+ """
47
+ if isinstance(value, torch.Tensor):
48
+ value = value.detach().cpu().numpy()
49
+
50
+ value = value.squeeze()
51
+ if invalid_mask is None:
52
+ invalid_mask = value == invalid_val
53
+ mask = np.logical_not(invalid_mask)
54
+
55
+ # normalize
56
+ vmin = np.percentile(value[mask],2) if vmin is None else vmin
57
+ vmax = np.percentile(value[mask],85) if vmax is None else vmax
58
+ if vmin != vmax:
59
+ value = (value - vmin) / (vmax - vmin) # vmin..vmax
60
+ else:
61
+ # Avoid 0-division
62
+ value = value * 0.
63
+
64
+ # squeeze last dim if it exists
65
+ # grey out the invalid values
66
+
67
+ value[invalid_mask] = np.nan
68
+ cmapper = matplotlib.cm.get_cmap(cmap)
69
+ if value_transform:
70
+ value = value_transform(value)
71
+ # value = value / value.max()
72
+ value = cmapper(value, bytes=True) # (nxmx4)
73
+
74
+ # img = value[:, :, :]
75
+ img = value[...]
76
+ img[invalid_mask] = background_color
77
+
78
+ # return img.transpose((2, 0, 1))
79
+ if gamma_corrected:
80
+ # gamma correction
81
+ img = img / 255
82
+ img = np.power(img, 2.2)
83
+ img = img * 255
84
+ img = img.astype(np.uint8)
85
+ return img
86
+