Bingxin commited on
Commit
7d8126a
·
verified ·
1 Parent(s): 7629caa

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +9 -19
README.md CHANGED
@@ -15,7 +15,8 @@ pinned: true
15
 
16
  <h1 align="center">Marigold Disparity v0.1 Model Card</h1>
17
 
18
- <!-- This model represents the internal test checkpoint of affine-invariant disparity version (train_marigold_affine_disparity_iter_24000). -->
 
19
 
20
 
21
  <p align="center">
@@ -34,9 +35,9 @@ pinned: true
34
  <a title="arXiv" href="https://arxiv.org/abs/2505.09358" target="_blank" rel="noopener noreferrer" style="display: inline-block;">
35
  <img src="https://img.shields.io/badge/%F0%9F%93%84%20Read%20-Paper-AF3436" alt="arXiv">
36
  </a>
37
- <a title="Social" href="https://twitter.com/antonobukhov1" target="_blank" rel="noopener noreferrer" style="display: inline-block;">
38
  <img src="https://img.shields.io/twitter/follow/:?label=Subscribe%20for%20updates!" alt="Social">
39
- </a>
40
  <a title="License" href="https://huggingface.co/stabilityai/stable-diffusion-2/blob/main/LICENSE-MODEL" target="_blank" rel="noopener noreferrer" style="display: inline-block;">
41
  <img src="https://img.shields.io/badge/License-OpenRAIL++-929292" alt="License">
42
  </a>
@@ -44,28 +45,17 @@ pinned: true
44
 
45
  This is a model card for the `marigold-disparity-affine-v0-1` model for monocular depth estimation from a single image.
46
  The model is fine-tuned from the `stable-diffusion-2` [model](https://huggingface.co/stabilityai/stable-diffusion-2) as
47
- described in our papers, in inverse depth (disparity) space:
48
  - [CVPR'2024 paper](https://hf.co/papers/2312.02145) titled "Repurposing Diffusion-Based Image Generators for Monocular Depth Estimation"
49
  - [Journal extension](https://hf.co/papers/2505.09358) titled "Marigold: Affordable Adaptation of Diffusion-Based Image Generators for Image Analysis"
50
 
51
  ### Using the model
52
- - Play with the interactive [Hugging Face Spaces demo](https://huggingface.co/spaces/prs-eth/marigold): check out how the model works with example images or upload your own.
53
- - Use it with [diffusers](https://huggingface.co/docs/diffusers/using-diffusers/marigold_usage) to compute the results with a few lines of code.
54
  - Get to the bottom of things with our [official codebase](https://github.com/prs-eth/marigold).
55
-
56
- ## Model Details
57
- - **Developed by:** [Bingxin Ke](http://www.kebingxin.com/), [Kevin Qu](https://ch.linkedin.com/in/kevin-qu-b3417621b), [Tianfu Wang](https://tianfwang.github.io/), [Nando Metzger](https://nandometzger.github.io/), [Shengyu Huang](https://shengyuh.github.io/), [Bo Li](https://www.linkedin.com/in/bobboli0202), [Anton Obukhov](https://www.obukhov.ai/), [Konrad Schindler](https://scholar.google.com/citations?user=FZuNgqIAAAAJ).
58
- - **Model type:** Generative latent diffusion-based affine-invariant monocular depth estimation from a single image.
59
  - **Language:** English.
60
- - **License:** [CreativeML Open RAIL++-M License](https://huggingface.co/stabilityai/stable-diffusion-2/blob/main/LICENSE-MODEL).
61
- - **Model Description:** This model can be used to generate an estimated depth map of an input image.
62
- - **Resolution**: Even though any resolution can be processed, the model inherits the base diffusion model's effective resolution of roughly **768** pixels.
63
- This means that for optimal predictions, any larger input image should be resized to make the longer side 768 pixels before feeding it into the model.
64
- - **Steps and scheduler**: This model was designed for usage with the **DDIM** scheduler and between **1 and 50** denoising steps.
65
- - **Outputs**:
66
- - **Affine-invariant depth map**: The predicted values are between 0 and 1, interpolating between the near and far planes of the model's choice.
67
- - **Uncertainty map**: Produced only when multiple predictions are ensembled with ensemble size larger than 2.
68
- - **Resources for more information:** [Project Website](https://marigoldcomputervision.github.io/), [Paper](https://arxiv.org/abs/2505.09358), [Code](https://github.com/prs-eth/marigold).
69
  - **Cite as:**
70
 
71
  ```bibtex
 
15
 
16
  <h1 align="center">Marigold Disparity v0.1 Model Card</h1>
17
 
18
+ This model is the internal test checkpoint of affine-invariant disparity version
19
+ <!-- (train_marigold_affine_disparity_iter_24000). -->
20
 
21
 
22
  <p align="center">
 
35
  <a title="arXiv" href="https://arxiv.org/abs/2505.09358" target="_blank" rel="noopener noreferrer" style="display: inline-block;">
36
  <img src="https://img.shields.io/badge/%F0%9F%93%84%20Read%20-Paper-AF3436" alt="arXiv">
37
  </a>
38
+ <!-- <a title="Social" href="https://twitter.com/antonobukhov1" target="_blank" rel="noopener noreferrer" style="display: inline-block;">
39
  <img src="https://img.shields.io/twitter/follow/:?label=Subscribe%20for%20updates!" alt="Social">
40
+ </a> -->
41
  <a title="License" href="https://huggingface.co/stabilityai/stable-diffusion-2/blob/main/LICENSE-MODEL" target="_blank" rel="noopener noreferrer" style="display: inline-block;">
42
  <img src="https://img.shields.io/badge/License-OpenRAIL++-929292" alt="License">
43
  </a>
 
45
 
46
  This is a model card for the `marigold-disparity-affine-v0-1` model for monocular depth estimation from a single image.
47
  The model is fine-tuned from the `stable-diffusion-2` [model](https://huggingface.co/stabilityai/stable-diffusion-2) as
48
+ described in our papers, in **inverse depth (disparity) space**:
49
  - [CVPR'2024 paper](https://hf.co/papers/2312.02145) titled "Repurposing Diffusion-Based Image Generators for Monocular Depth Estimation"
50
  - [Journal extension](https://hf.co/papers/2505.09358) titled "Marigold: Affordable Adaptation of Diffusion-Based Image Generators for Image Analysis"
51
 
52
  ### Using the model
53
+ - This model is for internal test.
 
54
  - Get to the bottom of things with our [official codebase](https://github.com/prs-eth/marigold).
55
+ - **Developed by:** [Bingxin Ke](http://www.kebingxin.com/), [Anton Obukhov](https://www.obukhov.ai/), [Shengyu Huang](https://shengyuh.github.io/), [Nando Metzger](https://nandometzger.github.io/), [Rodrigo Caye Daudt](https://rcdaudt.github.io/), [Konrad Schindler](https://scholar.google.com/citations?user=FZuNgqIAAAAJ).
56
+ - **Model type:** Generative latent diffusion-based affine-invariant disparity (inverse depth) estimation from a single image.
 
 
57
  - **Language:** English.
58
+ - **License:** [Apache License License Version 2.0](https://www.apache.org/licenses/LICENSE-2.0).
 
 
 
 
 
 
 
 
59
  - **Cite as:**
60
 
61
  ```bibtex