Depth Estimation¶
Class: DepthEstimationBlockV1
Source: inference.core.workflows.core_steps.models.foundation.depth_estimation.v1.DepthEstimationBlockV1
π― This workflow block performs depth estimation on images using Apple's DepthPro model. It analyzes the spatial relationships
and depth information in images to create a depth map where:
π Each pixel's value represents its relative distance from the camera
π Lower values (darker colors) indicate closer objects
π Higher values (lighter colors) indicate further objects
The model outputs:
1. πΊοΈ A depth map showing the relative distances of objects in the scene
2. π The camera's field of view (in degrees)
3. π¬ The camera's focal length
This is particularly useful for:
- ποΈ Understanding 3D structure from 2D images
- π¨ Creating depth-aware visualizations
- π Analyzing spatial relationships in scenes
- πΆοΈ Applications in augmented reality and 3D reconstruction
β‘ The model runs efficiently on Apple Silicon (M1-M4) using Metal Performance Shaders (MPS) for accelerated inference.
Type identifier¶
Use the following identifier in step "type" field: roboflow_core/depth_estimation@v1to add the block as
as step in your workflow.
Properties¶
| Name | Type | Description | Refs |
|---|---|---|---|
name |
str |
Enter a unique identifier for this step.. | β |
model_version |
str |
The Depth Estimation model to be used for inference.. | β |
The Refs column marks possibility to parametrise the property with dynamic values available
in workflow runtime. See Bindings for more info.
Available Connections¶
Compatible Blocks
Check what blocks you can connect to Depth Estimation in version v1.
- inputs:
Stitch Images,Image Threshold,Email Notification,Corner Visualization,Image Blur,Ellipse Visualization,OpenAI,Roboflow Dataset Upload,Object Detection Model,Depth Estimation,Stitch OCR Detections,EasyOCR,Absolute Static Crop,CogVLM,Google Gemini,Stability AI Image Generation,Grid Visualization,Dynamic Crop,Image Slicer,Image Preprocessing,Relative Static Crop,SIFT,Morphological Transformation,Instance Segmentation Model,Line Counter Visualization,Trace Visualization,LMM For Classification,Halo Visualization,Dot Visualization,GLM-OCR,Model Monitoring Inference Aggregator,Roboflow Custom Metadata,Pixelate Visualization,Circle Visualization,Image Convert Grayscale,Icon Visualization,QR Code Generator,S3 Sink,Keypoint Detection Model,Twilio SMS Notification,Halo Visualization,Camera Focus,Anthropic Claude,OCR Model,Polygon Visualization,Text Display,Reference Path Visualization,Llama 3.2 Vision,CSV Formatter,Crop Visualization,Roboflow Dataset Upload,Mask Visualization,Heatmap Visualization,Webhook Sink,Label Visualization,Classification Label Visualization,Google Vision OCR,Florence-2 Model,Florence-2 Model,VLM As Detector,Polygon Zone Visualization,Stability AI Inpainting,Google Gemini,Perspective Correction,Camera Calibration,Anthropic Claude,OpenAI,OpenAI,Qwen3.5-VL,Background Color Visualization,Anthropic Claude,Email Notification,Background Subtraction,Contrast Equalization,SIFT Comparison,Multi-Label Classification Model,Keypoint Visualization,Stitch OCR Detections,LMM,Color Visualization,Single-Label Classification Model,OpenAI,Roboflow Vision Events,Local File Sink,VLM As Classifier,Twilio SMS/MMS Notification,Triangle Visualization,Clip Comparison,Blur Visualization,Bounding Box Visualization,Camera Focus,Polygon Visualization,Google Gemini,Image Slicer,Image Contours,Model Comparison Visualization,Stability AI Outpainting,Slack Notification - outputs:
Stitch Images,Image Threshold,Corner Visualization,Image Blur,Ellipse Visualization,OpenAI,Roboflow Dataset Upload,Object Detection Model,Barcode Detection,Depth Estimation,Gaze Detection,EasyOCR,Absolute Static Crop,Multi-Label Classification Model,CogVLM,Google Gemini,Stability AI Image Generation,Time in Zone,Dynamic Crop,Instance Segmentation Model,Image Slicer,Image Preprocessing,SIFT,Relative Static Crop,Morphological Transformation,Line Counter Visualization,Trace Visualization,LMM For Classification,Halo Visualization,Dot Visualization,ByteTrack Tracker,GLM-OCR,Keypoint Detection Model,Pixel Color Count,Pixelate Visualization,Circle Visualization,Semantic Segmentation Model,Image Convert Grayscale,Icon Visualization,Keypoint Detection Model,Detections Stabilizer,Halo Visualization,Camera Focus,SIFT Comparison,Anthropic Claude,SAM 3,OC-SORT Tracker,OCR Model,Polygon Visualization,Text Display,Qwen2.5-VL,Reference Path Visualization,Instance Segmentation Model,Llama 3.2 Vision,Qwen3-VL,Crop Visualization,Roboflow Dataset Upload,Mask Visualization,Moondream2,CLIP Embedding Model,SORT Tracker,Heatmap Visualization,Label Visualization,Google Vision OCR,Byte Tracker,Classification Label Visualization,Florence-2 Model,Segment Anything 2 Model,Florence-2 Model,VLM As Detector,Polygon Zone Visualization,Stability AI Inpainting,VLM As Classifier,Google Gemini,SAM 3,Perspective Correction,Anthropic Claude,OpenAI,Camera Calibration,VLM As Detector,OpenAI,Qwen3.5-VL,Template Matching,Background Color Visualization,Anthropic Claude,Email Notification,Background Subtraction,SIFT Comparison,Contrast Equalization,Multi-Label Classification Model,Keypoint Visualization,LMM,SmolVLM2,Single-Label Classification Model,Perception Encoder Embedding Model,SAM 3,Motion Detection,Seg Preview,Single-Label Classification Model,Color Visualization,Object Detection Model,OpenAI,Dominant Color,Roboflow Vision Events,QR Code Detection,VLM As Classifier,Detections Stitch,Clip Comparison,YOLO-World Model,Triangle Visualization,Buffer,Clip Comparison,Twilio SMS/MMS Notification,Bounding Box Visualization,Blur Visualization,Camera Focus,Polygon Visualization,Google Gemini,Image Slicer,Image Contours,Model Comparison Visualization,Stability AI Outpainting
Input and Output Bindings¶
The available connections depend on its binding kinds. Check what binding kinds
Depth Estimation in version v1 has.
Bindings
-
input
-
output
image(image): Image in workflows.normalized_depth(numpy_array): Numpy array.
Example JSON definition of step Depth Estimation in version v1
{
"name": "<your_step_name_here>",
"type": "roboflow_core/depth_estimation@v1",
"images": "$inputs.image",
"model_version": "depth-anything-v2/small"
}