Towards Robust Monocular Depth Estimation: Mixing Datasets for Zero-shot Cross-dataset Transfer
Paper
•
1907.01341
•
Published
DPT 3.1 (MiDaS) models, leveraging state-of-the-art vision backbones such as BEiT and Swinv2
Note This model gives the highest quality, but is also the most heavy in terms of computation as mentioned in the paper.
Note This model has moderately less quality, but has a better speed-performance trade-off
Note This model is recommended for deployment on embedded devices