Depth Estimation¶
Class: DepthEstimationBlockV1
Source: inference.core.workflows.core_steps.models.foundation.depth_estimation.v1.DepthEstimationBlockV1
π― This workflow block performs depth estimation on images using Apple's DepthPro model. It analyzes the spatial relationships
and depth information in images to create a depth map where:
π Each pixel's value represents its relative distance from the camera
π Lower values (darker colors) indicate closer objects
π Higher values (lighter colors) indicate further objects
The model outputs:
1. πΊοΈ A depth map showing the relative distances of objects in the scene
2. π The camera's field of view (in degrees)
3. π¬ The camera's focal length
This is particularly useful for:
- ποΈ Understanding 3D structure from 2D images
- π¨ Creating depth-aware visualizations
- π Analyzing spatial relationships in scenes
- πΆοΈ Applications in augmented reality and 3D reconstruction
β‘ The model runs efficiently on Apple Silicon (M1-M4) using Metal Performance Shaders (MPS) for accelerated inference.
Type identifier¶
Use the following identifier in step "type" field: roboflow_core/depth_estimation@v1to add the block as
as step in your workflow.
Properties¶
| Name | Type | Description | Refs |
|---|---|---|---|
name |
str |
Enter a unique identifier for this step.. | β |
model_version |
str |
The Depth Estimation model to be used for inference.. | β |
The Refs column marks possibility to parametrise the property with dynamic values available
in workflow runtime. See Bindings for more info.
Available Connections¶
Compatible Blocks
Check what blocks you can connect to Depth Estimation in version v1.
- inputs:
QR Code Generator,Image Convert Grayscale,Google Gemini,Dynamic Crop,Blur Visualization,SIFT,Bounding Box Visualization,Stability AI Outpainting,Camera Focus,Slack Notification,Keypoint Visualization,Trace Visualization,Polygon Visualization,Ellipse Visualization,Model Comparison Visualization,OpenAI,Anthropic Claude,Local File Sink,Triangle Visualization,Polygon Zone Visualization,Halo Visualization,LMM,Stability AI Image Generation,Florence-2 Model,Circle Visualization,Email Notification,Google Vision OCR,Google Gemini,Clip Comparison,Camera Focus,Anthropic Claude,Object Detection Model,Instance Segmentation Model,Perspective Correction,CSV Formatter,Reference Path Visualization,Corner Visualization,Color Visualization,Twilio SMS/MMS Notification,VLM as Classifier,Image Slicer,OpenAI,Stitch OCR Detections,Camera Calibration,Image Blur,VLM as Detector,Dot Visualization,Roboflow Custom Metadata,Image Threshold,Model Monitoring Inference Aggregator,Morphological Transformation,Label Visualization,Background Color Visualization,Classification Label Visualization,OCR Model,Roboflow Dataset Upload,Mask Visualization,Pixelate Visualization,Absolute Static Crop,Keypoint Detection Model,Webhook Sink,Grid Visualization,Contrast Equalization,Image Preprocessing,Google Gemini,Relative Static Crop,Stability AI Inpainting,Image Contours,Line Counter Visualization,Stitch Images,OpenAI,Crop Visualization,OpenAI,Llama 3.2 Vision,Icon Visualization,SIFT Comparison,Depth Estimation,Twilio SMS Notification,Single-Label Classification Model,Florence-2 Model,Background Subtraction,LMM For Classification,Multi-Label Classification Model,EasyOCR,CogVLM,Image Slicer,Roboflow Dataset Upload,Email Notification - outputs:
Detections Stitch,Seg Preview,Byte Tracker,Qwen3-VL,YOLO-World Model,Google Gemini,Image Convert Grayscale,QR Code Detection,Dynamic Crop,SIFT Comparison,Blur Visualization,SIFT,Stability AI Outpainting,Bounding Box Visualization,Camera Focus,Keypoint Visualization,Trace Visualization,Instance Segmentation Model,Polygon Visualization,Dominant Color,Pixel Color Count,Ellipse Visualization,OpenAI,Model Comparison Visualization,Anthropic Claude,Triangle Visualization,Qwen2.5-VL,SAM 3,Polygon Zone Visualization,Halo Visualization,LMM,Stability AI Image Generation,Time in Zone,VLM as Detector,Florence-2 Model,CLIP Embedding Model,Single-Label Classification Model,Detections Stabilizer,Email Notification,Circle Visualization,Google Vision OCR,Google Gemini,Motion Detection,Clip Comparison,Camera Focus,Anthropic Claude,Object Detection Model,Instance Segmentation Model,Perspective Correction,Perception Encoder Embedding Model,VLM as Classifier,Reference Path Visualization,Corner Visualization,Color Visualization,Twilio SMS/MMS Notification,EasyOCR,Multi-Label Classification Model,Image Slicer,OpenAI,Image Blur,Buffer,Camera Calibration,SmolVLM2,VLM as Detector,SAM 3,Dot Visualization,Image Threshold,Morphological Transformation,Label Visualization,Background Color Visualization,OCR Model,Classification Label Visualization,Roboflow Dataset Upload,Keypoint Detection Model,Mask Visualization,Pixelate Visualization,Absolute Static Crop,Keypoint Detection Model,Moondream2,Image Preprocessing,Contrast Equalization,Google Gemini,Stability AI Inpainting,Barcode Detection,Template Matching,Line Counter Visualization,OpenAI,Image Contours,Crop Visualization,Relative Static Crop,Stitch Images,OpenAI,Llama 3.2 Vision,Icon Visualization,Clip Comparison,SIFT Comparison,Gaze Detection,Depth Estimation,Single-Label Classification Model,VLM as Classifier,Florence-2 Model,Background Subtraction,LMM For Classification,SAM 3,Object Detection Model,Multi-Label Classification Model,Segment Anything 2 Model,CogVLM,Image Slicer,Roboflow Dataset Upload
Input and Output Bindings¶
The available connections depend on its binding kinds. Check what binding kinds
Depth Estimation in version v1 has.
Bindings
-
input
-
output
image(image): Image in workflows.normalized_depth(numpy_array): Numpy array.
Example JSON definition of step Depth Estimation in version v1
{
"name": "<your_step_name_here>",
"type": "roboflow_core/depth_estimation@v1",
"images": "$inputs.image",
"model_version": "depth-anything-v2/small"
}