Depth Estimation¶
Class: DepthEstimationBlockV1
Source: inference.core.workflows.core_steps.models.foundation.depth_estimation.v1.DepthEstimationBlockV1
π― This workflow block performs depth estimation on images using Apple's DepthPro model. It analyzes the spatial relationships
and depth information in images to create a depth map where:
π Each pixel's value represents its relative distance from the camera
π Lower values (darker colors) indicate closer objects
π Higher values (lighter colors) indicate further objects
The model outputs:
1. πΊοΈ A depth map showing the relative distances of objects in the scene
2. π The camera's field of view (in degrees)
3. π¬ The camera's focal length
This is particularly useful for:
- ποΈ Understanding 3D structure from 2D images
- π¨ Creating depth-aware visualizations
- π Analyzing spatial relationships in scenes
- πΆοΈ Applications in augmented reality and 3D reconstruction
β‘ The model runs efficiently on Apple Silicon (M1-M4) using Metal Performance Shaders (MPS) for accelerated inference.
Type identifier¶
Use the following identifier in step "type" field: roboflow_core/depth_estimation@v1to add the block as
as step in your workflow.
Properties¶
| Name | Type | Description | Refs |
|---|---|---|---|
name |
str |
Enter a unique identifier for this step.. | β |
model_version |
str |
The Depth Estimation model to be used for inference.. | β |
The Refs column marks possibility to parametrise the property with dynamic values available
in workflow runtime. See Bindings for more info.
Available Connections¶
Compatible Blocks
Check what blocks you can connect to Depth Estimation in version v1.
- inputs:
Label Visualization,Blur Visualization,Background Color Visualization,Contrast Equalization,Bounding Box Visualization,Camera Calibration,Polygon Visualization,Stability AI Outpainting,Image Slicer,Keypoint Visualization,Reference Path Visualization,Pixelate Visualization,Icon Visualization,Triangle Visualization,Model Comparison Visualization,Corner Visualization,Image Preprocessing,Color Visualization,SIFT Comparison,Line Counter Visualization,Grid Visualization,Stitch Images,Halo Visualization,Stability AI Image Generation,QR Code Generator,Circle Visualization,Image Contours,Relative Static Crop,Dot Visualization,Polygon Zone Visualization,Ellipse Visualization,Image Blur,Absolute Static Crop,Depth Estimation,Image Slicer,Morphological Transformation,Stability AI Inpainting,Dynamic Crop,Camera Focus,Crop Visualization,Image Threshold,Perspective Correction,Image Convert Grayscale,Mask Visualization,Trace Visualization,Classification Label Visualization,SIFT - outputs:
Google Vision OCR,Label Visualization,LMM For Classification,Blur Visualization,Background Color Visualization,Contrast Equalization,Reference Path Visualization,Keypoint Visualization,Stability AI Outpainting,Bounding Box Visualization,Image Slicer,Pixelate Visualization,Single-Label Classification Model,Clip Comparison,SAM 3,Perception Encoder Embedding Model,Seg Preview,Byte Tracker,Image Preprocessing,SAM 3,Color Visualization,SIFT Comparison,Qwen2.5-VL,Object Detection Model,Dominant Color,Anthropic Claude,Circle Visualization,Image Contours,Object Detection Model,QR Code Detection,Polygon Zone Visualization,Ellipse Visualization,Email Notification,Clip Comparison,Moondream2,VLM as Classifier,OCR Model,Absolute Static Crop,Depth Estimation,LMM,Time in Zone,Morphological Transformation,Roboflow Dataset Upload,Gaze Detection,Crop Visualization,OpenAI,Florence-2 Model,Barcode Detection,Image Convert Grayscale,SAM 3,CogVLM,VLM as Detector,Multi-Label Classification Model,Classification Label Visualization,Buffer,Keypoint Detection Model,Segment Anything 2 Model,Keypoint Detection Model,SIFT Comparison,YOLO-World Model,Polygon Visualization,CLIP Embedding Model,Camera Calibration,Icon Visualization,Triangle Visualization,Template Matching,Roboflow Dataset Upload,Anthropic Claude,Model Comparison Visualization,Corner Visualization,Florence-2 Model,Google Gemini,Google Gemini,EasyOCR,VLM as Detector,Line Counter Visualization,SmolVLM2,Halo Visualization,Stability AI Image Generation,Relative Static Crop,Dot Visualization,Detections Stitch,Llama 3.2 Vision,Image Blur,OpenAI,Instance Segmentation Model,Multi-Label Classification Model,Image Slicer,OpenAI,Stability AI Inpainting,Dynamic Crop,Single-Label Classification Model,Camera Focus,Pixel Color Count,Detections Stabilizer,Instance Segmentation Model,VLM as Classifier,Mask Visualization,Perspective Correction,Image Threshold,OpenAI,Trace Visualization,Stitch Images,SIFT
Input and Output Bindings¶
The available connections depend on its binding kinds. Check what binding kinds
Depth Estimation in version v1 has.
Bindings
-
input
images(image): The image to infer on..
-
output
image(image): Image in workflows.normalized_depth(numpy_array): Numpy array.
Example JSON definition of step Depth Estimation in version v1
{
"name": "<your_step_name_here>",
"type": "roboflow_core/depth_estimation@v1",
"images": "$inputs.image",
"model_version": "depth-anything-v2/small"
}