Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
@@ -30,7 +30,7 @@ model.load_state_dict(torch.load('checkpoints/depth_anything_vitl14.pth'))
|
|
30 |
|
31 |
title = "# Depth Anything"
|
32 |
description = """Official demo for **Depth Anything: Unleashing the Power of Large-Scale Unlabeled Data**.
|
33 |
-
Please refer to our [paper](), [project page](https://depth-anything.github.io), or [github](https://github.com/LiheYoung/Depth-Anything) for more details."""
|
34 |
|
35 |
transform = Compose([
|
36 |
Resize(
|
@@ -56,6 +56,7 @@ with gr.Blocks(css=css) as demo:
|
|
56 |
gr.Markdown(title)
|
57 |
gr.Markdown(description)
|
58 |
gr.Markdown("### Depth Prediction demo")
|
|
|
59 |
|
60 |
with gr.Row():
|
61 |
input_image = gr.Image(label="Input Image", type='numpy', elem_id='img-display-input')
|
|
|
30 |
|
31 |
title = "# Depth Anything"
|
32 |
description = """Official demo for **Depth Anything: Unleashing the Power of Large-Scale Unlabeled Data**.
|
33 |
+
Please refer to our [paper](https://arxiv.org/abs/2401.10891), [project page](https://depth-anything.github.io), or [github](https://github.com/LiheYoung/Depth-Anything) for more details."""
|
34 |
|
35 |
transform = Compose([
|
36 |
Resize(
|
|
|
56 |
gr.Markdown(title)
|
57 |
gr.Markdown(description)
|
58 |
gr.Markdown("### Depth Prediction demo")
|
59 |
+
gr.Markdown("You can slide the output to compare the depth prediction with input image")
|
60 |
|
61 |
with gr.Row():
|
62 |
input_image = gr.Image(label="Input Image", type='numpy', elem_id='img-display-input')
|