LiheYoung commited on
Commit
f24c91c
1 Parent(s): 3930ec3

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +2 -1
app.py CHANGED
@@ -30,7 +30,7 @@ model.load_state_dict(torch.load('checkpoints/depth_anything_vitl14.pth'))
30
 
31
  title = "# Depth Anything"
32
  description = """Official demo for **Depth Anything: Unleashing the Power of Large-Scale Unlabeled Data**.
33
- Please refer to our [paper](), [project page](https://depth-anything.github.io), or [github](https://github.com/LiheYoung/Depth-Anything) for more details."""
34
 
35
  transform = Compose([
36
  Resize(
@@ -56,6 +56,7 @@ with gr.Blocks(css=css) as demo:
56
  gr.Markdown(title)
57
  gr.Markdown(description)
58
  gr.Markdown("### Depth Prediction demo")
 
59
 
60
  with gr.Row():
61
  input_image = gr.Image(label="Input Image", type='numpy', elem_id='img-display-input')
 
30
 
31
  title = "# Depth Anything"
32
  description = """Official demo for **Depth Anything: Unleashing the Power of Large-Scale Unlabeled Data**.
33
+ Please refer to our [paper](https://arxiv.org/abs/2401.10891), [project page](https://depth-anything.github.io), or [github](https://github.com/LiheYoung/Depth-Anything) for more details."""
34
 
35
  transform = Compose([
36
  Resize(
 
56
  gr.Markdown(title)
57
  gr.Markdown(description)
58
  gr.Markdown("### Depth Prediction demo")
59
+ gr.Markdown("You can slide the output to compare the depth prediction with input image")
60
 
61
  with gr.Row():
62
  input_image = gr.Image(label="Input Image", type='numpy', elem_id='img-display-input')