SAM 3¶
v3¶
Class: SegmentAnything3BlockV3 (there are multiple versions of this block)
Source: inference.core.workflows.core_steps.models.foundation.segment_anything3.v3.SegmentAnything3BlockV3
Warning: This block has multiple versions. Please refer to the specific version for details. You can learn more about how versions work here: Versioning
Run Segment Anything 3 (SAM3), a zero-shot instance segmentation model, on an image.
You can use text prompts for open-vocabulary segmentation - just specify class names and SAM3 will segment those objects in the image.
This block supports two output formats: - rle (default): Returns masks in RLE (Run-Length Encoding) format, which is more memory-efficient - polygons: Returns polygon coordinates for each mask
RLE format is recommended for high-resolution images or workflows with many detections.
Type identifier¶
Use the following identifier in step "type" field: roboflow_core/sam3@v3to add the block as
as step in your workflow.
Properties¶
| Name | Type | Description | Refs |
|---|---|---|---|
name |
str |
Enter a unique identifier for this step.. | ❌ |
model_id |
str |
model version. You only need to change this for fine tuned sam3 models.. | ✅ |
class_names |
Optional[List[str], str] |
List of classes to recognise. | ✅ |
confidence |
float |
Minimum confidence threshold for predicted masks. | ✅ |
per_class_confidence |
List[float] |
List of confidence thresholds per class (must match class_names length). | ✅ |
apply_nms |
bool |
Whether to apply Non-Maximum Suppression across prompts. | ✅ |
nms_iou_threshold |
float |
IoU threshold for cross-prompt NMS. Must be in [0.0, 1.0]. | ✅ |
output_format |
str |
'rle' returns efficient RLE encoding (recommended), 'polygons' returns polygon coordinates. | ❌ |
The Refs column marks possibility to parametrise the property with dynamic values available
in workflow runtime. See Bindings for more info.
Available Connections¶
Compatible Blocks
Check what blocks you can connect to SAM 3 in version v3.
- inputs:
Dynamic Crop,OCR Model,Motion Detection,Email Notification,Image Blur,Background Subtraction,SIFT Comparison,OpenAI,Google Vision OCR,Google Gemini,Image Preprocessing,Instance Segmentation Model,Object Detection Model,Local File Sink,Single-Label Classification Model,Bounding Box Visualization,Multi-Label Classification Model,Model Monitoring Inference Aggregator,Anthropic Claude,Multi-Label Classification Model,Keypoint Detection Model,Email Notification,Slack Notification,Camera Focus,Twilio SMS/MMS Notification,Identify Outliers,VLM As Detector,Dot Visualization,Florence-2 Model,Gaze Detection,Roboflow Dataset Upload,CSV Formatter,Camera Focus,Stitch OCR Detections,Depth Estimation,Polygon Visualization,Perspective Correction,OpenAI,PTZ Tracking (ONVIF),Camera Calibration,Corner Visualization,Icon Visualization,Image Slicer,Qwen3.5-VL,Line Counter Visualization,Heatmap Visualization,Morphological Transformation,Stability AI Image Generation,Detections List Roll-Up,Google Gemini,Keypoint Visualization,VLM As Detector,Halo Visualization,Keypoint Detection Model,Background Color Visualization,Label Visualization,JSON Parser,Polygon Visualization,Pixelate Visualization,LMM,CogVLM,Single-Label Classification Model,Contrast Equalization,Triangle Visualization,Stability AI Outpainting,Dimension Collapse,Mask Visualization,VLM As Classifier,Color Visualization,Instance Segmentation Model,Text Display,Relative Static Crop,Reference Path Visualization,Stitch OCR Detections,Llama 3.2 Vision,OpenAI,Image Threshold,Clip Comparison,Classification Label Visualization,Clip Comparison,Webhook Sink,Circle Visualization,Polygon Zone Visualization,Image Contours,Image Convert Grayscale,Grid Visualization,VLM As Classifier,Florence-2 Model,Buffer,Roboflow Custom Metadata,LMM For Classification,Dynamic Zone,SIFT,Halo Visualization,Object Detection Model,Semantic Segmentation Model,Detections Consensus,Anthropic Claude,Google Gemini,Model Comparison Visualization,Blur Visualization,QR Code Generator,EasyOCR,Absolute Static Crop,Image Slicer,S3 Sink,Anthropic Claude,Cosine Similarity,Stability AI Inpainting,Ellipse Visualization,Identify Changes,Crop Visualization,SIFT Comparison,Trace Visualization,Twilio SMS Notification,Stitch Images,Size Measurement,OpenAI,Roboflow Dataset Upload - outputs:
Dynamic Crop,Time in Zone,Circle Visualization,Time in Zone,Mask Area Measurement,Byte Tracker,Bounding Box Visualization,Model Monitoring Inference Aggregator,Florence-2 Model,Detections Stitch,Roboflow Custom Metadata,Dot Visualization,Florence-2 Model,Dynamic Zone,Roboflow Dataset Upload,Camera Focus,Velocity,Polygon Visualization,Perspective Correction,PTZ Tracking (ONVIF),Halo Visualization,Line Counter,Icon Visualization,Corner Visualization,Overlap Filter,Detections List Roll-Up,Byte Tracker,Detections Consensus,Heatmap Visualization,Distance Measurement,Byte Tracker,Model Comparison Visualization,Blur Visualization,Detection Offset,Halo Visualization,Background Color Visualization,Path Deviation,Label Visualization,Detection Event Log,Detections Merge,Polygon Visualization,Pixelate Visualization,Time in Zone,Triangle Visualization,Detections Transformation,Mask Visualization,Ellipse Visualization,Stability AI Inpainting,Crop Visualization,Path Deviation,Trace Visualization,Color Visualization,Detections Combine,Detections Classes Replacement,Segment Anything 2 Model,Size Measurement,Bounding Rectangle,Line Counter,Detections Stabilizer,Detections Filter,Roboflow Dataset Upload
Input and Output Bindings¶
The available connections depend on its binding kinds. Check what binding kinds
SAM 3 in version v3 has.
Bindings
-
input
images(image): The image to infer on..model_id(roboflow_model_id): model version. You only need to change this for fine tuned sam3 models..class_names(Union[string,list_of_values]): List of classes to recognise.confidence(float): Minimum confidence threshold for predicted masks.per_class_confidence(list_of_values): List of confidence thresholds per class (must match class_names length).apply_nms(boolean): Whether to apply Non-Maximum Suppression across prompts.nms_iou_threshold(float): IoU threshold for cross-prompt NMS. Must be in [0.0, 1.0].
-
output
predictions(Union[rle_instance_segmentation_prediction,instance_segmentation_prediction]): Prediction with detected bounding boxes and RLE-encoded segmentation masks in form of sv.Detections(...) object ifrle_instance_segmentation_predictionor Prediction with detected bounding boxes and segmentation masks in form of sv.Detections(...) object ifinstance_segmentation_prediction.
Example JSON definition of step SAM 3 in version v3
{
"name": "<your_step_name_here>",
"type": "roboflow_core/sam3@v3",
"images": "$inputs.image",
"model_id": "sam3/sam3_final",
"class_names": [
"car",
"person"
],
"confidence": 0.3,
"per_class_confidence": [
0.3,
0.5,
0.7
],
"apply_nms": "<block_does_not_provide_example>",
"nms_iou_threshold": 0.5,
"output_format": "rle"
}
v2¶
Class: SegmentAnything3BlockV2 (there are multiple versions of this block)
Source: inference.core.workflows.core_steps.models.foundation.segment_anything3.v2.SegmentAnything3BlockV2
Warning: This block has multiple versions. Please refer to the specific version for details. You can learn more about how versions work here: Versioning
Run Segment Anything 3, a zero-shot instance segmentation model, on an image.
You can pass in boxes/predictions from other models as prompts, or use a text prompt for open-vocabulary segmentation. If you pass in box detections from another model, the class names of the boxes will be forwarded to the predicted masks.
Type identifier¶
Use the following identifier in step "type" field: roboflow_core/sam3@v2to add the block as
as step in your workflow.
Properties¶
| Name | Type | Description | Refs |
|---|---|---|---|
name |
str |
Enter a unique identifier for this step.. | ❌ |
model_id |
str |
model version. You only need to change this for fine tuned sam3 models.. | ✅ |
class_names |
Optional[List[str], str] |
List of classes to recognise. | ✅ |
confidence |
float |
Minimum confidence threshold for predicted masks. | ✅ |
per_class_confidence |
List[float] |
List of confidence thresholds per class (must match class_names length). | ✅ |
apply_nms |
bool |
Whether to apply Non-Maximum Suppression across prompts. | ✅ |
nms_iou_threshold |
float |
IoU threshold for cross-prompt NMS. Must be in [0.0, 1.0]. | ✅ |
The Refs column marks possibility to parametrise the property with dynamic values available
in workflow runtime. See Bindings for more info.
Available Connections¶
Compatible Blocks
Check what blocks you can connect to SAM 3 in version v2.
- inputs:
Dynamic Crop,OCR Model,Motion Detection,Email Notification,Image Blur,Background Subtraction,SIFT Comparison,OpenAI,Google Vision OCR,Google Gemini,Image Preprocessing,Instance Segmentation Model,Object Detection Model,Local File Sink,Single-Label Classification Model,Bounding Box Visualization,Multi-Label Classification Model,Model Monitoring Inference Aggregator,Anthropic Claude,Multi-Label Classification Model,Keypoint Detection Model,Email Notification,Slack Notification,Camera Focus,Twilio SMS/MMS Notification,Identify Outliers,VLM As Detector,Dot Visualization,Florence-2 Model,Gaze Detection,Roboflow Dataset Upload,CSV Formatter,Camera Focus,Stitch OCR Detections,Depth Estimation,Polygon Visualization,Perspective Correction,OpenAI,PTZ Tracking (ONVIF),Camera Calibration,Corner Visualization,Icon Visualization,Image Slicer,Qwen3.5-VL,Line Counter Visualization,Heatmap Visualization,Morphological Transformation,Stability AI Image Generation,Detections List Roll-Up,Google Gemini,Keypoint Visualization,VLM As Detector,Halo Visualization,Keypoint Detection Model,Background Color Visualization,Label Visualization,JSON Parser,Polygon Visualization,Pixelate Visualization,LMM,CogVLM,Single-Label Classification Model,Contrast Equalization,Triangle Visualization,Stability AI Outpainting,Dimension Collapse,Mask Visualization,VLM As Classifier,Color Visualization,Instance Segmentation Model,Text Display,Relative Static Crop,Reference Path Visualization,Stitch OCR Detections,Llama 3.2 Vision,OpenAI,Image Threshold,Clip Comparison,Classification Label Visualization,Clip Comparison,Webhook Sink,Circle Visualization,Polygon Zone Visualization,Image Contours,Image Convert Grayscale,Grid Visualization,VLM As Classifier,Florence-2 Model,Buffer,Roboflow Custom Metadata,LMM For Classification,Dynamic Zone,SIFT,Halo Visualization,Object Detection Model,Semantic Segmentation Model,Detections Consensus,Anthropic Claude,Google Gemini,Model Comparison Visualization,Blur Visualization,QR Code Generator,EasyOCR,Absolute Static Crop,Image Slicer,S3 Sink,Anthropic Claude,Cosine Similarity,Stability AI Inpainting,Ellipse Visualization,Identify Changes,Crop Visualization,SIFT Comparison,Trace Visualization,Twilio SMS Notification,Stitch Images,Size Measurement,OpenAI,Roboflow Dataset Upload - outputs:
Dynamic Crop,Time in Zone,Circle Visualization,Time in Zone,Mask Area Measurement,Byte Tracker,Model Monitoring Inference Aggregator,Bounding Box Visualization,Florence-2 Model,Detections Stitch,Roboflow Custom Metadata,Dot Visualization,Florence-2 Model,Dynamic Zone,Roboflow Dataset Upload,Camera Focus,Velocity,Polygon Visualization,Halo Visualization,PTZ Tracking (ONVIF),Perspective Correction,Line Counter,Icon Visualization,Corner Visualization,Overlap Filter,Byte Tracker,Detections Consensus,Detections List Roll-Up,Heatmap Visualization,Distance Measurement,Byte Tracker,Model Comparison Visualization,Blur Visualization,Detection Offset,Halo Visualization,Background Color Visualization,Path Deviation,Label Visualization,Detection Event Log,Polygon Visualization,Pixelate Visualization,Time in Zone,Triangle Visualization,Detections Transformation,Mask Visualization,Ellipse Visualization,Path Deviation,Crop Visualization,Stability AI Inpainting,Trace Visualization,Color Visualization,Segment Anything 2 Model,Detections Combine,Size Measurement,Detections Stabilizer,Detections Filter,Bounding Rectangle,Detections Classes Replacement,Detections Merge,Roboflow Dataset Upload,Line Counter
Input and Output Bindings¶
The available connections depend on its binding kinds. Check what binding kinds
SAM 3 in version v2 has.
Bindings
-
input
images(image): The image to infer on..model_id(roboflow_model_id): model version. You only need to change this for fine tuned sam3 models..class_names(Union[string,list_of_values]): List of classes to recognise.confidence(float): Minimum confidence threshold for predicted masks.per_class_confidence(list_of_values): List of confidence thresholds per class (must match class_names length).apply_nms(boolean): Whether to apply Non-Maximum Suppression across prompts.nms_iou_threshold(float): IoU threshold for cross-prompt NMS. Must be in [0.0, 1.0].
-
output
predictions(instance_segmentation_prediction): Prediction with detected bounding boxes and segmentation masks in form of sv.Detections(...) object.
Example JSON definition of step SAM 3 in version v2
{
"name": "<your_step_name_here>",
"type": "roboflow_core/sam3@v2",
"images": "$inputs.image",
"model_id": "sam3/sam3_final",
"class_names": [
"car",
"person"
],
"confidence": 0.3,
"per_class_confidence": [
0.3,
0.5,
0.7
],
"apply_nms": "<block_does_not_provide_example>",
"nms_iou_threshold": 0.5
}
v1¶
Class: SegmentAnything3BlockV1 (there are multiple versions of this block)
Source: inference.core.workflows.core_steps.models.foundation.segment_anything3.v1.SegmentAnything3BlockV1
Warning: This block has multiple versions. Please refer to the specific version for details. You can learn more about how versions work here: Versioning
Run Segment Anything 3, a zero-shot instance segmentation model, on an image.
You can pass in boxes/predictions from other models as prompts, or use a text prompt for open-vocabulary segmentation. If you pass in box detections from another model, the class names of the boxes will be forwarded to the predicted masks.
Type identifier¶
Use the following identifier in step "type" field: roboflow_core/sam3@v1to add the block as
as step in your workflow.
Properties¶
| Name | Type | Description | Refs |
|---|---|---|---|
name |
str |
Enter a unique identifier for this step.. | ❌ |
model_id |
str |
model version. You only need to change this for fine tuned sam3 models.. | ✅ |
class_names |
Optional[List[str], str] |
List of classes to recognise. | ✅ |
threshold |
float |
Threshold for predicted mask scores. | ✅ |
The Refs column marks possibility to parametrise the property with dynamic values available
in workflow runtime. See Bindings for more info.
Available Connections¶
Compatible Blocks
Check what blocks you can connect to SAM 3 in version v1.
- inputs:
Dynamic Crop,OCR Model,Motion Detection,Email Notification,Image Blur,Background Subtraction,SIFT Comparison,OpenAI,Google Vision OCR,Google Gemini,Image Preprocessing,Instance Segmentation Model,Object Detection Model,Local File Sink,Single-Label Classification Model,Bounding Box Visualization,Multi-Label Classification Model,Model Monitoring Inference Aggregator,Anthropic Claude,Multi-Label Classification Model,Keypoint Detection Model,Email Notification,Slack Notification,Camera Focus,Twilio SMS/MMS Notification,Dot Visualization,Florence-2 Model,Gaze Detection,Roboflow Dataset Upload,CSV Formatter,Camera Focus,Stitch OCR Detections,Depth Estimation,Polygon Visualization,Perspective Correction,OpenAI,Camera Calibration,Corner Visualization,Icon Visualization,Image Slicer,Qwen3.5-VL,Line Counter Visualization,Heatmap Visualization,Morphological Transformation,Stability AI Image Generation,Detections List Roll-Up,Google Gemini,Keypoint Visualization,VLM As Detector,Halo Visualization,Keypoint Detection Model,Background Color Visualization,Label Visualization,Polygon Visualization,Pixelate Visualization,LMM,CogVLM,Single-Label Classification Model,Contrast Equalization,Triangle Visualization,Stability AI Outpainting,Dimension Collapse,Mask Visualization,VLM As Classifier,Color Visualization,Instance Segmentation Model,Text Display,Relative Static Crop,Reference Path Visualization,Stitch OCR Detections,Llama 3.2 Vision,OpenAI,Image Threshold,Clip Comparison,Classification Label Visualization,Clip Comparison,Webhook Sink,Circle Visualization,Polygon Zone Visualization,Image Contours,Image Convert Grayscale,Grid Visualization,Florence-2 Model,Buffer,Roboflow Custom Metadata,LMM For Classification,Dynamic Zone,SIFT,Halo Visualization,Object Detection Model,Semantic Segmentation Model,Anthropic Claude,Google Gemini,Model Comparison Visualization,Blur Visualization,QR Code Generator,EasyOCR,Absolute Static Crop,Image Slicer,S3 Sink,Anthropic Claude,Cosine Similarity,Stability AI Inpainting,Ellipse Visualization,Identify Changes,Crop Visualization,Trace Visualization,Twilio SMS Notification,Stitch Images,Size Measurement,OpenAI,Roboflow Dataset Upload - outputs:
Dynamic Crop,Time in Zone,Circle Visualization,Time in Zone,Mask Area Measurement,Byte Tracker,Model Monitoring Inference Aggregator,Bounding Box Visualization,Florence-2 Model,Detections Stitch,Roboflow Custom Metadata,Dot Visualization,Florence-2 Model,Dynamic Zone,Roboflow Dataset Upload,Camera Focus,Velocity,Polygon Visualization,Halo Visualization,PTZ Tracking (ONVIF),Perspective Correction,Line Counter,Icon Visualization,Corner Visualization,Overlap Filter,Byte Tracker,Detections Consensus,Detections List Roll-Up,Heatmap Visualization,Distance Measurement,Byte Tracker,Model Comparison Visualization,Blur Visualization,Detection Offset,Halo Visualization,Background Color Visualization,Path Deviation,Label Visualization,Detection Event Log,Polygon Visualization,Pixelate Visualization,Time in Zone,Triangle Visualization,Detections Transformation,Mask Visualization,Ellipse Visualization,Path Deviation,Crop Visualization,Stability AI Inpainting,Trace Visualization,Color Visualization,Segment Anything 2 Model,Detections Combine,Size Measurement,Detections Stabilizer,Detections Filter,Bounding Rectangle,Detections Classes Replacement,Detections Merge,Roboflow Dataset Upload,Line Counter
Input and Output Bindings¶
The available connections depend on its binding kinds. Check what binding kinds
SAM 3 in version v1 has.
Bindings
-
input
images(image): The image to infer on..model_id(roboflow_model_id): model version. You only need to change this for fine tuned sam3 models..class_names(Union[string,list_of_values]): List of classes to recognise.threshold(float): Threshold for predicted mask scores.
-
output
predictions(instance_segmentation_prediction): Prediction with detected bounding boxes and segmentation masks in form of sv.Detections(...) object.
Example JSON definition of step SAM 3 in version v1
{
"name": "<your_step_name_here>",
"type": "roboflow_core/sam3@v1",
"images": "$inputs.image",
"model_id": "sam3/sam3_final",
"class_names": [
"car",
"person"
],
"threshold": 0.3
}