Update README.md
Browse files
README.md
CHANGED
@@ -86,7 +86,7 @@ post_processed_output = image_processor.post_process_depth_estimation(
|
|
86 |
field_of_view = post_processed_output[0]["field_of_view"]
|
87 |
focal_length = post_processed_output[0]["focal_length"]
|
88 |
depth = post_processed_output[0]["predicted_depth"]
|
89 |
-
depth = (depth - depth.min()) / depth.max()
|
90 |
depth = depth * 255.
|
91 |
depth = depth.detach().cpu().numpy()
|
92 |
depth = Image.fromarray(depth.astype("uint8"))
|
@@ -131,7 +131,7 @@ The `DepthProEncoder` further uses two encoders:
|
|
131 |
- `image_encoder`
|
132 |
- Input image is also rescaled to `patch_size` and processed by the **`image_encoder`**
|
133 |
|
134 |
-
Both these encoders can be configured via `patch_model_config` and `image_model_config` respectively, both of which are
|
135 |
|
136 |
Outputs from both encoders (`last_hidden_state`) and selected intermediate states (`hidden_states`) from **`patch_encoder`** are fused by a `DPT`-based `FeatureFusionStage` for depth estimation.
|
137 |
|
|
|
86 |
field_of_view = post_processed_output[0]["field_of_view"]
|
87 |
focal_length = post_processed_output[0]["focal_length"]
|
88 |
depth = post_processed_output[0]["predicted_depth"]
|
89 |
+
depth = (depth - depth.min()) / (depth.max() - depth.min())
|
90 |
depth = depth * 255.
|
91 |
depth = depth.detach().cpu().numpy()
|
92 |
depth = Image.fromarray(depth.astype("uint8"))
|
|
|
131 |
- `image_encoder`
|
132 |
- Input image is also rescaled to `patch_size` and processed by the **`image_encoder`**
|
133 |
|
134 |
+
Both these encoders can be configured via `patch_model_config` and `image_model_config` respectively, both of which are separate `Dinov2Model` by default.
|
135 |
|
136 |
Outputs from both encoders (`last_hidden_state`) and selected intermediate states (`hidden_states`) from **`patch_encoder`** are fused by a `DPT`-based `FeatureFusionStage` for depth estimation.
|
137 |
|