|
--- |
|
license: mit |
|
tags: |
|
- depth-estimation |
|
- onnx |
|
- computer-vision |
|
- visiondepth3d |
|
- mit-license |
|
--- |
|
|
|
# Distill-Any-Depth-Large (ONNX) – For VisionDepth3D |
|
|
|
> **Model Origin:** This model is based on [Distill-Any-Depth by Westlake-AGI-Lab](https://github.com/Westlake-AGI-Lab/Distill-Any-Depth), originally developed by Westlake-AGI-Lab. |
|
> I did not train this model — I have converted it to ONNX format for fast, GPU-accelerated inference within tools such as VisionDepth3D. |
|
|
|
## 🧠 About This Model |
|
|
|
This is a direct conversion of the **Distill-Any-Depth** PyTorch model to **ONNX**, real-time depth estimation from single RGB images. |
|
|
|
### ✔️ Key Features: |
|
- ONNX format (exported from PyTorch) |
|
- Compatible with ONNX Runtime and TensorRT |
|
- Excellent for 2D to 3D depth workflows |
|
- Works seamlessly with **VisionDepth3D** |
|
|
|
## 📌 Intended Use |
|
|
|
- Real-time or batch depth map generation |
|
- 2D to 3D conversion pipelines (e.g., SBS 3D video) |
|
- Works on Windows, Linux (CUDA-supported) |
|
|
|
## 📜 License and Attribution |
|
|
|
### Citation |
|
|
|
``` |
|
@article{he2025distill, |
|
title = {Distill Any Depth: Distillation Creates a Stronger Monocular Depth Estimator}, |
|
author = {Xiankang He and Dongyan Guo and Hongji Li and Ruibo Li and Ying Cui and Chi Zhang}, |
|
year = {2025}, |
|
journal = {arXiv preprint arXiv: 2502.19204} |
|
} |
|
``` |
|
|
|
- **Source Model:** [Distill-Any-Depth by Westlake-AGI-Lab](https://github.com/Westlake-AGI-Lab/Distill-Any-Depth) |
|
- **License:** MIT |
|
- **Modifications:** Only format conversion (no retraining or weight changes) |
|
|
|
> If you use this model, please credit the original authors: Westlake-AGI-Lab. |
|
|
|
## 💻 How to Use In VisionDepth3D |
|
|
|
Place Folder containing onnx model into weights folder in VisionDepth3D |
|
|
|
``` |
|
VisionDepth3D¬ |
|
Weights¬ |
|
Distill Any Depth Large¬ |
|
model.onnx |
|
``` |