Upgraded-Depth-Anything-V2 - UDAV2 extension for a1111. It significantly outperforms V1 in fine-grained details & robustness. Compared with SD-based models, it enjoys faster inference speed, fewer parameters & higher depth accuracy.
- 2024-06-14: Paper, project page, code, models, demo, & benchmark are all released.
- 2024-06-20: The repo has been upgraded & is also now running on .safetensors models instead of .pth models.
- 2024-06-23: Updated installation process to be a simpler one_click_install.bat file. It automatically downloads the depth models into a 'checkpoints' folder, the triton wheel into the repo's main folder & installs all of the dependencies needed. [Also updated this README.md file to provide more clarity!] Upgraded-Depth-Anything-V2 - UDAV2
- 2024-06-24: pravdomil has provided a much need update to UDAV2 for 16bit image creation in order to make stunning 3D Bas-Reliefs! I am currently in the process of updating the gradio webui to include both 16bit single image & 16bit batch image creation which will be pushed in the coming days.
- 2024-06-25: Working on a beta version of UDAV2 as an a1111 extension & will be released next week, so stay-tuned.
- 2024-06-27: A1111 extension released!
- 2024-06-29: Updated forge extension release sd-forge-udav2, to prevent conflicts w/ pre-existing installed extensions in Forge! OUTDATED
- 2024-07-01: Extension has been added to the extension index.json! You can now install this extension directly inside A1111.
- 2024-07-03: Released v0.0.3 of sd-forge-udav2, meant for anyone using the forge-webui!
- 2024-07-03: [v1.1.452] sd-webui-controlnet now has a depth_anything_v2 preprocessor🔥! Update transformers dependency to transformers-4.44.1 to use the new depth_anything_v2 controlnet preprocessor.
- 2024-07-24: Updated upgraded_depth_anything_v2.py script to fix install & usage errors.
- 2024-08-02: Released v0.0.5 of sd-forge-udav2, updated to prevent depth_anything_v2.dpt NoModule & gradio integration errors w/ Forge.
All you need to do is clone this repo into the stable-diffusion-webui/extensions
folder & then run a1111. Or You can install the extension directly from the a1111 extensions tab now!
git clone https://github.com/MackinationsAi/sd-webui-udav2.git
If you want to use this extension w/ stable-diffusion-forge-webui
please install this extension sd-forge-udav2, NOT sd-webui-udav2
.
- This extension has basically all of the same features & functionality as the main UDAV2 stand-alone repo w/ some minor changes for the conversion to a1111 extension to work properly.
- The 'Today's Depth Gallery' tab automatically updates after each single image or batch image processing for the current day.
- All of the outputs are saved in a very similar structure to 'txt2img' & 'img2img' in the main a1111 outputs folder under 'depths'
- I'm still making updates to the main which I will push to this extension when possible! (Including 16bit for hires 3D depth mapping)
- Bug fixed -
It usually took between 45-90 secs the first time you launch the webui after installing the extension (as it is downloading the models, once everything is installed you will see in cmd that everything has been installed & will progress w/ the launch).
Lihe Yang1 · Bingyi Kang2† · Zilong Huang2 · Zhen Zhao · Xiaogang Xu · Jiashi Feng2 · Hengshuang Zhao1*
Legend Keys - [ HKU 1 · TikTok 2 · project-lead † · corresponding author * ]
Depth-Anything-V2-Small model is under the Apache-2.0 license. Depth-Anything-V2-Base/Large/Giant models are under the CC-BY-NC-4.0 license.
If you find this project useful, please consider citing below, give this upgraded repo a star & share it w/ others in the community!
@article{depth_anything_v2,
title={Depth Anything V2},
author={Yang, Lihe & Kang, Bingyi & Huang, Zilong & Zhao, Zhen & Xu, Xiaogang & Feng, Jiashi & Zhao, Hengshuang},
journal={arXiv:2406.09414},
year={2024}
}
@inproceedings{depth_anything_v1,
title={Depth Anything: Unleashing the Power of Large-Scale Unlabeled Data},
author={Yang, Lihe & Kang, Bingyi & Huang, Zilong & Xu, Xiaogang & Feng, Jiashi & Zhao, Hengshuang},
booktitle={CVPR},
year={2024}
}