Depth Estimation¶
Class: DepthEstimationBlockV1
Source: inference.core.workflows.core_steps.models.foundation.depth_estimation.v1.DepthEstimationBlockV1
π― This workflow block performs depth estimation on images using Apple's DepthPro model. It analyzes the spatial relationships
and depth information in images to create a depth map where:
π Each pixel's value represents its relative distance from the camera
π Lower values (darker colors) indicate closer objects
π Higher values (lighter colors) indicate further objects
The model outputs:
1. πΊοΈ A depth map showing the relative distances of objects in the scene
2. π The camera's field of view (in degrees)
3. π¬ The camera's focal length
This is particularly useful for:
- ποΈ Understanding 3D structure from 2D images
- π¨ Creating depth-aware visualizations
- π Analyzing spatial relationships in scenes
- πΆοΈ Applications in augmented reality and 3D reconstruction
β‘ The model runs efficiently on Apple Silicon (M1-M4) using Metal Performance Shaders (MPS) for accelerated inference.
Type identifier¶
Use the following identifier in step "type" field: roboflow_core/depth_estimation@v1to add the block as
as step in your workflow.
Properties¶
| Name | Type | Description | Refs |
|---|---|---|---|
name |
str |
Enter a unique identifier for this step.. | β |
model_version |
str |
The Depth Estimation model to be used for inference.. | β |
The Refs column marks possibility to parametrise the property with dynamic values available
in workflow runtime. See Bindings for more info.
Available Connections¶
Compatible Blocks
Check what blocks you can connect to Depth Estimation in version v1.
- inputs:
Single-Label Classification Model,OpenAI,OpenAI,Slack Notification,Anthropic Claude,SIFT,Halo Visualization,Google Gemini,Camera Focus,Image Convert Grayscale,Image Blur,LMM For Classification,VLM As Classifier,Google Gemini,Mask Visualization,Anthropic Claude,Polygon Visualization,LMM,Email Notification,Dynamic Crop,Camera Calibration,Color Visualization,SIFT Comparison,Background Subtraction,Grid Visualization,Polygon Visualization,Image Contours,VLM As Detector,Polygon Zone Visualization,Twilio SMS Notification,Webhook Sink,Roboflow Vision Events,Email Notification,Absolute Static Crop,Roboflow Custom Metadata,Label Visualization,OpenAI,Image Slicer,Florence-2 Model,Twilio SMS/MMS Notification,Circle Visualization,Object Detection Model,Local File Sink,Stability AI Outpainting,Reference Path Visualization,Keypoint Detection Model,Llama 3.2 Vision,EasyOCR,Crop Visualization,CogVLM,Dot Visualization,Roboflow Dataset Upload,Stitch OCR Detections,Bounding Box Visualization,Model Monitoring Inference Aggregator,Heatmap Visualization,Icon Visualization,Image Preprocessing,Stability AI Image Generation,Stability AI Inpainting,Anthropic Claude,Pixelate Visualization,Relative Static Crop,Camera Focus,Ellipse Visualization,Background Color Visualization,Qwen3.5-VL,OCR Model,CSV Formatter,Blur Visualization,Stitch OCR Detections,Keypoint Visualization,Corner Visualization,Stitch Images,Morphological Transformation,Clip Comparison,GLM-OCR,Model Comparison Visualization,Contrast Equalization,Perspective Correction,Triangle Visualization,Image Threshold,Halo Visualization,Trace Visualization,Classification Label Visualization,Florence-2 Model,Line Counter Visualization,Google Vision OCR,Google Gemini,Roboflow Dataset Upload,Multi-Label Classification Model,OpenAI,Image Slicer,S3 Sink,Instance Segmentation Model,QR Code Generator,Text Display,Depth Estimation - outputs:
Single-Label Classification Model,OpenAI,Anthropic Claude,SIFT,Multi-Label Classification Model,Halo Visualization,Google Gemini,SAM2 Video Tracker,Camera Focus,Barcode Detection,Image Convert Grayscale,SIFT Comparison,VLM As Detector,LMM For Classification,Template Matching,Qwen3-VL,Google Gemini,OC-SORT Tracker,LMM,Color Visualization,Camera Calibration,SIFT Comparison,Background Subtraction,Polygon Visualization,Image Contours,Polygon Zone Visualization,Roboflow Vision Events,Email Notification,Absolute Static Crop,SAM 3,Label Visualization,Florence-2 Model,YOLO-World Model,Segment Anything 2 Model,SAM 3,Llama 3.2 Vision,Object Detection Model,EasyOCR,Instance Segmentation Model,Crop Visualization,Roboflow Dataset Upload,Bounding Box Visualization,Single-Label Classification Model,Heatmap Visualization,Icon Visualization,Anthropic Claude,Pixelate Visualization,Ellipse Visualization,Background Color Visualization,Camera Focus,Qwen3.5-VL,Buffer,Relative Static Crop,Blur Visualization,Object Detection Model,Stitch Images,Dominant Color,GLM-OCR,Contrast Equalization,Keypoint Detection Model,Triangle Visualization,Perspective Correction,Time in Zone,Image Threshold,Halo Visualization,Trace Visualization,Classification Label Visualization,SmolVLM2,Florence-2 Model,Line Counter Visualization,OpenAI,VLM As Classifier,Seg Preview,Instance Segmentation Model,Keypoint Detection Model,Text Display,Single-Label Classification Model,OpenAI,Gaze Detection,SORT Tracker,Image Blur,ByteTrack Tracker,Instance Segmentation Model,VLM As Classifier,Multi-Label Classification Model,Mask Visualization,Anthropic Claude,Polygon Visualization,Qwen2.5-VL,Dynamic Crop,Moondream2,VLM As Detector,Semantic Segmentation Model,Byte Tracker,OpenAI,Image Slicer,Twilio SMS/MMS Notification,Object Detection Model,Circle Visualization,CLIP Embedding Model,Stability AI Outpainting,Keypoint Detection Model,Reference Path Visualization,CogVLM,Dot Visualization,Pixel Color Count,Image Preprocessing,Clip Comparison,Stability AI Image Generation,Stability AI Inpainting,OCR Model,Keypoint Visualization,Detections Stabilizer,Semantic Segmentation Model,Corner Visualization,Motion Detection,Morphological Transformation,Clip Comparison,Model Comparison Visualization,QR Code Detection,Roboflow Dataset Upload,Perception Encoder Embedding Model,Google Vision OCR,Google Gemini,Multi-Label Classification Model,Image Slicer,Detections Stitch,SAM 3,Depth Estimation
Input and Output Bindings¶
The available connections depend on its binding kinds. Check what binding kinds
Depth Estimation in version v1 has.
Bindings
-
input
-
output
image(image): Image in workflows.normalized_depth(numpy_array): Numpy array.
Example JSON definition of step Depth Estimation in version v1
{
"name": "<your_step_name_here>",
"type": "roboflow_core/depth_estimation@v1",
"images": "$inputs.image",
"model_version": "depth-anything-v2/small"
}