SAM 3¶
v3¶
Class: SegmentAnything3BlockV3 (there are multiple versions of this block)
Source: inference.core.workflows.core_steps.models.foundation.segment_anything3.v3.SegmentAnything3BlockV3
Warning: This block has multiple versions. Please refer to the specific version for details. You can learn more about how versions work here: Versioning
Run Segment Anything 3 (SAM3), a zero-shot instance segmentation model, on an image.
You can use text prompts for open-vocabulary segmentation - just specify class names and SAM3 will segment those objects in the image.
This block supports two output formats: - rle (default): Returns masks in RLE (Run-Length Encoding) format, which is more memory-efficient - polygons: Returns polygon coordinates for each mask
RLE format is recommended for high-resolution images or workflows with many detections.
Type identifier¶
Use the following identifier in step "type" field: roboflow_core/sam3@v3to add the block as
as step in your workflow.
Properties¶
| Name | Type | Description | Refs |
|---|---|---|---|
name |
str |
Enter a unique identifier for this step.. | ❌ |
model_id |
str |
model version. You only need to change this for fine tuned sam3 models.. | ✅ |
class_names |
Optional[List[str], str] |
List of classes to recognise. | ✅ |
confidence |
float |
Minimum confidence threshold for predicted masks. | ✅ |
per_class_confidence |
List[float] |
List of confidence thresholds per class (must match class_names length). | ✅ |
apply_nms |
bool |
Whether to apply Non-Maximum Suppression across prompts. | ✅ |
nms_iou_threshold |
float |
IoU threshold for cross-prompt NMS. Must be in [0.0, 1.0]. | ✅ |
output_format |
str |
'rle' returns efficient RLE encoding (recommended), 'polygons' returns polygon coordinates. | ❌ |
The Refs column marks possibility to parametrise the property with dynamic values available
in workflow runtime. See Bindings for more info.
Available Connections¶
Compatible Blocks
Check what blocks you can connect to SAM 3 in version v3.
- inputs:
Clip Comparison,Halo Visualization,Anthropic Claude,Image Blur,Email Notification,Camera Focus,Text Display,CSV Formatter,Contrast Equalization,Object Detection Model,Blur Visualization,VLM As Detector,Roboflow Dataset Upload,Corner Visualization,Dynamic Crop,Classification Label Visualization,Relative Static Crop,Dimension Collapse,Trace Visualization,Keypoint Detection Model,Multi-Label Classification Model,Camera Focus,Mask Visualization,SIFT,Stitch OCR Detections,VLM As Classifier,Background Color Visualization,PTZ Tracking (ONVIF).md),Polygon Visualization,Local File Sink,Florence-2 Model,Stitch OCR Detections,Size Measurement,Ellipse Visualization,Pixelate Visualization,SIFT Comparison,Roboflow Custom Metadata,Single-Label Classification Model,Identify Changes,LMM,Detections List Roll-Up,OpenAI,Circle Visualization,Dot Visualization,Twilio SMS Notification,Identify Outliers,SIFT Comparison,Absolute Static Crop,Morphological Transformation,Crop Visualization,Single-Label Classification Model,Dynamic Zone,Google Gemini,Keypoint Visualization,Polygon Visualization,Google Vision OCR,Icon Visualization,Multi-Label Classification Model,Llama 3.2 Vision,Gaze Detection,Color Visualization,Image Contours,Stitch Images,OpenAI,VLM As Classifier,LMM For Classification,Email Notification,Cosine Similarity,VLM As Detector,Anthropic Claude,OpenAI,Bounding Box Visualization,Triangle Visualization,Background Subtraction,Grid Visualization,Model Comparison Visualization,Reference Path Visualization,Line Counter Visualization,EasyOCR,Halo Visualization,OCR Model,Instance Segmentation Model,Slack Notification,Stability AI Inpainting,Detections Consensus,Image Slicer,Google Gemini,Florence-2 Model,JSON Parser,Image Convert Grayscale,Stability AI Image Generation,Heatmap Visualization,Polygon Zone Visualization,Image Slicer,Label Visualization,Google Gemini,Depth Estimation,Image Preprocessing,OpenAI,Object Detection Model,Stability AI Outpainting,CogVLM,Webhook Sink,Image Threshold,Instance Segmentation Model,Anthropic Claude,Buffer,Camera Calibration,Perspective Correction,QR Code Generator,Motion Detection,Keypoint Detection Model,Model Monitoring Inference Aggregator,Twilio SMS/MMS Notification,Roboflow Dataset Upload,Clip Comparison - outputs:
Polygon Visualization,Halo Visualization,Icon Visualization,Color Visualization,Segment Anything 2 Model,Blur Visualization,Detections Stabilizer,Roboflow Dataset Upload,Corner Visualization,Dynamic Crop,Distance Measurement,Trace Visualization,Triangle Visualization,Bounding Box Visualization,Detections Classes Replacement,Model Comparison Visualization,Mask Visualization,Camera Focus,Halo Visualization,Detections Consensus,Time in Zone,Stability AI Inpainting,Path Deviation,Line Counter,Detections Combine,Detection Event Log,Florence-2 Model,Background Color Visualization,Mask Area Measurement,Bounding Rectangle,PTZ Tracking (ONVIF).md),Heatmap Visualization,Polygon Visualization,Florence-2 Model,Size Measurement,Detections Filter,Line Counter,Ellipse Visualization,Label Visualization,Pixelate Visualization,Time in Zone,Roboflow Custom Metadata,Byte Tracker,Detections Stitch,Time in Zone,Detections List Roll-Up,Detections Merge,Dot Visualization,Circle Visualization,Byte Tracker,Byte Tracker,Path Deviation,Velocity,Perspective Correction,Detections Transformation,Overlap Filter,Model Monitoring Inference Aggregator,Detection Offset,Crop Visualization,Roboflow Dataset Upload,Dynamic Zone
Input and Output Bindings¶
The available connections depend on its binding kinds. Check what binding kinds
SAM 3 in version v3 has.
Bindings
-
input
images(image): The image to infer on..model_id(roboflow_model_id): model version. You only need to change this for fine tuned sam3 models..class_names(Union[string,list_of_values]): List of classes to recognise.confidence(float): Minimum confidence threshold for predicted masks.per_class_confidence(list_of_values): List of confidence thresholds per class (must match class_names length).apply_nms(boolean): Whether to apply Non-Maximum Suppression across prompts.nms_iou_threshold(float): IoU threshold for cross-prompt NMS. Must be in [0.0, 1.0].
-
output
predictions(Union[rle_instance_segmentation_prediction,instance_segmentation_prediction]): Prediction with detected bounding boxes and RLE-encoded segmentation masks in form of sv.Detections(...) object ifrle_instance_segmentation_predictionor Prediction with detected bounding boxes and segmentation masks in form of sv.Detections(...) object ifinstance_segmentation_prediction.
Example JSON definition of step SAM 3 in version v3
{
"name": "<your_step_name_here>",
"type": "roboflow_core/sam3@v3",
"images": "$inputs.image",
"model_id": "sam3/sam3_final",
"class_names": [
"car",
"person"
],
"confidence": 0.3,
"per_class_confidence": [
0.3,
0.5,
0.7
],
"apply_nms": "<block_does_not_provide_example>",
"nms_iou_threshold": 0.5,
"output_format": "rle"
}
v2¶
Class: SegmentAnything3BlockV2 (there are multiple versions of this block)
Source: inference.core.workflows.core_steps.models.foundation.segment_anything3.v2.SegmentAnything3BlockV2
Warning: This block has multiple versions. Please refer to the specific version for details. You can learn more about how versions work here: Versioning
Run Segment Anything 3, a zero-shot instance segmentation model, on an image.
You can pass in boxes/predictions from other models as prompts, or use a text prompt for open-vocabulary segmentation. If you pass in box detections from another model, the class names of the boxes will be forwarded to the predicted masks.
Type identifier¶
Use the following identifier in step "type" field: roboflow_core/sam3@v2to add the block as
as step in your workflow.
Properties¶
| Name | Type | Description | Refs |
|---|---|---|---|
name |
str |
Enter a unique identifier for this step.. | ❌ |
model_id |
str |
model version. You only need to change this for fine tuned sam3 models.. | ✅ |
class_names |
Optional[List[str], str] |
List of classes to recognise. | ✅ |
confidence |
float |
Minimum confidence threshold for predicted masks. | ✅ |
per_class_confidence |
List[float] |
List of confidence thresholds per class (must match class_names length). | ✅ |
apply_nms |
bool |
Whether to apply Non-Maximum Suppression across prompts. | ✅ |
nms_iou_threshold |
float |
IoU threshold for cross-prompt NMS. Must be in [0.0, 1.0]. | ✅ |
The Refs column marks possibility to parametrise the property with dynamic values available
in workflow runtime. See Bindings for more info.
Available Connections¶
Compatible Blocks
Check what blocks you can connect to SAM 3 in version v2.
- inputs:
Clip Comparison,Halo Visualization,Anthropic Claude,Image Blur,Email Notification,Camera Focus,Text Display,CSV Formatter,Contrast Equalization,Object Detection Model,Blur Visualization,VLM As Detector,Roboflow Dataset Upload,Corner Visualization,Dynamic Crop,Classification Label Visualization,Relative Static Crop,Dimension Collapse,Trace Visualization,Keypoint Detection Model,Multi-Label Classification Model,Camera Focus,Mask Visualization,SIFT,Stitch OCR Detections,VLM As Classifier,Background Color Visualization,PTZ Tracking (ONVIF).md),Polygon Visualization,Local File Sink,Florence-2 Model,Stitch OCR Detections,Size Measurement,Ellipse Visualization,Pixelate Visualization,SIFT Comparison,Roboflow Custom Metadata,Single-Label Classification Model,Identify Changes,LMM,Detections List Roll-Up,OpenAI,Circle Visualization,Dot Visualization,Twilio SMS Notification,Identify Outliers,SIFT Comparison,Absolute Static Crop,Morphological Transformation,Crop Visualization,Single-Label Classification Model,Dynamic Zone,Google Gemini,Keypoint Visualization,Polygon Visualization,Google Vision OCR,Icon Visualization,Multi-Label Classification Model,Llama 3.2 Vision,Gaze Detection,Color Visualization,Image Contours,Stitch Images,OpenAI,VLM As Classifier,LMM For Classification,Email Notification,Cosine Similarity,VLM As Detector,Anthropic Claude,OpenAI,Bounding Box Visualization,Triangle Visualization,Background Subtraction,Grid Visualization,Model Comparison Visualization,Reference Path Visualization,Line Counter Visualization,EasyOCR,Halo Visualization,OCR Model,Instance Segmentation Model,Slack Notification,Stability AI Inpainting,Detections Consensus,Image Slicer,Google Gemini,Florence-2 Model,JSON Parser,Image Convert Grayscale,Stability AI Image Generation,Heatmap Visualization,Polygon Zone Visualization,Image Slicer,Label Visualization,Google Gemini,Depth Estimation,Image Preprocessing,OpenAI,Object Detection Model,Stability AI Outpainting,CogVLM,Webhook Sink,Image Threshold,Instance Segmentation Model,Anthropic Claude,Buffer,Camera Calibration,Perspective Correction,QR Code Generator,Motion Detection,Keypoint Detection Model,Model Monitoring Inference Aggregator,Twilio SMS/MMS Notification,Roboflow Dataset Upload,Clip Comparison - outputs:
Polygon Visualization,Halo Visualization,Icon Visualization,Color Visualization,Segment Anything 2 Model,Blur Visualization,Detections Stabilizer,Roboflow Dataset Upload,Corner Visualization,Dynamic Crop,Distance Measurement,Trace Visualization,Triangle Visualization,Detections Classes Replacement,Bounding Box Visualization,Model Comparison Visualization,Mask Visualization,Camera Focus,Halo Visualization,Detections Consensus,Time in Zone,Stability AI Inpainting,Line Counter,Detection Event Log,Detections Combine,Path Deviation,Florence-2 Model,Background Color Visualization,Mask Area Measurement,Bounding Rectangle,PTZ Tracking (ONVIF).md),Heatmap Visualization,Polygon Visualization,Florence-2 Model,Size Measurement,Label Visualization,Line Counter,Ellipse Visualization,Detections Filter,Pixelate Visualization,Time in Zone,Roboflow Custom Metadata,Byte Tracker,Detections Stitch,Time in Zone,Detections List Roll-Up,Detections Merge,Dot Visualization,Circle Visualization,Byte Tracker,Byte Tracker,Path Deviation,Velocity,Perspective Correction,Detections Transformation,Overlap Filter,Model Monitoring Inference Aggregator,Detection Offset,Crop Visualization,Roboflow Dataset Upload,Dynamic Zone
Input and Output Bindings¶
The available connections depend on its binding kinds. Check what binding kinds
SAM 3 in version v2 has.
Bindings
-
input
images(image): The image to infer on..model_id(roboflow_model_id): model version. You only need to change this for fine tuned sam3 models..class_names(Union[string,list_of_values]): List of classes to recognise.confidence(float): Minimum confidence threshold for predicted masks.per_class_confidence(list_of_values): List of confidence thresholds per class (must match class_names length).apply_nms(boolean): Whether to apply Non-Maximum Suppression across prompts.nms_iou_threshold(float): IoU threshold for cross-prompt NMS. Must be in [0.0, 1.0].
-
output
predictions(instance_segmentation_prediction): Prediction with detected bounding boxes and segmentation masks in form of sv.Detections(...) object.
Example JSON definition of step SAM 3 in version v2
{
"name": "<your_step_name_here>",
"type": "roboflow_core/sam3@v2",
"images": "$inputs.image",
"model_id": "sam3/sam3_final",
"class_names": [
"car",
"person"
],
"confidence": 0.3,
"per_class_confidence": [
0.3,
0.5,
0.7
],
"apply_nms": "<block_does_not_provide_example>",
"nms_iou_threshold": 0.5
}
v1¶
Class: SegmentAnything3BlockV1 (there are multiple versions of this block)
Source: inference.core.workflows.core_steps.models.foundation.segment_anything3.v1.SegmentAnything3BlockV1
Warning: This block has multiple versions. Please refer to the specific version for details. You can learn more about how versions work here: Versioning
Run Segment Anything 3, a zero-shot instance segmentation model, on an image.
You can pass in boxes/predictions from other models as prompts, or use a text prompt for open-vocabulary segmentation. If you pass in box detections from another model, the class names of the boxes will be forwarded to the predicted masks.
Type identifier¶
Use the following identifier in step "type" field: roboflow_core/sam3@v1to add the block as
as step in your workflow.
Properties¶
| Name | Type | Description | Refs |
|---|---|---|---|
name |
str |
Enter a unique identifier for this step.. | ❌ |
model_id |
str |
model version. You only need to change this for fine tuned sam3 models.. | ✅ |
class_names |
Optional[List[str], str] |
List of classes to recognise. | ✅ |
threshold |
float |
Threshold for predicted mask scores. | ✅ |
The Refs column marks possibility to parametrise the property with dynamic values available
in workflow runtime. See Bindings for more info.
Available Connections¶
Compatible Blocks
Check what blocks you can connect to SAM 3 in version v1.
- inputs:
Clip Comparison,Halo Visualization,Anthropic Claude,Image Blur,Email Notification,Camera Focus,Text Display,CSV Formatter,Contrast Equalization,Object Detection Model,Blur Visualization,Roboflow Dataset Upload,Corner Visualization,Dynamic Crop,Classification Label Visualization,Relative Static Crop,Dimension Collapse,Trace Visualization,Keypoint Detection Model,Multi-Label Classification Model,Camera Focus,Mask Visualization,SIFT,Stitch OCR Detections,Background Color Visualization,Polygon Visualization,Local File Sink,Florence-2 Model,Stitch OCR Detections,Size Measurement,Ellipse Visualization,Pixelate Visualization,SIFT Comparison,Roboflow Custom Metadata,Single-Label Classification Model,Identify Changes,LMM,Detections List Roll-Up,OpenAI,Circle Visualization,Dot Visualization,Twilio SMS Notification,Absolute Static Crop,Morphological Transformation,Crop Visualization,Single-Label Classification Model,Dynamic Zone,Google Gemini,Keypoint Visualization,Polygon Visualization,Google Vision OCR,Icon Visualization,Multi-Label Classification Model,Llama 3.2 Vision,Gaze Detection,Color Visualization,Image Contours,Stitch Images,OpenAI,VLM As Classifier,LMM For Classification,Email Notification,Cosine Similarity,VLM As Detector,Anthropic Claude,OpenAI,Bounding Box Visualization,Triangle Visualization,Background Subtraction,Grid Visualization,Model Comparison Visualization,Reference Path Visualization,Line Counter Visualization,EasyOCR,Halo Visualization,OCR Model,Instance Segmentation Model,Slack Notification,Stability AI Inpainting,Image Slicer,Google Gemini,Florence-2 Model,Image Convert Grayscale,Stability AI Image Generation,Heatmap Visualization,Polygon Zone Visualization,Image Slicer,Label Visualization,Google Gemini,Depth Estimation,Image Preprocessing,OpenAI,Object Detection Model,Stability AI Outpainting,CogVLM,Webhook Sink,Image Threshold,Instance Segmentation Model,Anthropic Claude,Buffer,Camera Calibration,Perspective Correction,QR Code Generator,Motion Detection,Keypoint Detection Model,Model Monitoring Inference Aggregator,Twilio SMS/MMS Notification,Roboflow Dataset Upload,Clip Comparison - outputs:
Polygon Visualization,Halo Visualization,Icon Visualization,Color Visualization,Segment Anything 2 Model,Blur Visualization,Detections Stabilizer,Roboflow Dataset Upload,Corner Visualization,Dynamic Crop,Distance Measurement,Trace Visualization,Triangle Visualization,Detections Classes Replacement,Bounding Box Visualization,Model Comparison Visualization,Mask Visualization,Camera Focus,Halo Visualization,Detections Consensus,Time in Zone,Stability AI Inpainting,Line Counter,Detection Event Log,Detections Combine,Path Deviation,Florence-2 Model,Background Color Visualization,Mask Area Measurement,Bounding Rectangle,PTZ Tracking (ONVIF).md),Heatmap Visualization,Polygon Visualization,Florence-2 Model,Size Measurement,Label Visualization,Line Counter,Ellipse Visualization,Detections Filter,Pixelate Visualization,Time in Zone,Roboflow Custom Metadata,Byte Tracker,Detections Stitch,Time in Zone,Detections List Roll-Up,Detections Merge,Dot Visualization,Circle Visualization,Byte Tracker,Byte Tracker,Path Deviation,Velocity,Perspective Correction,Detections Transformation,Overlap Filter,Model Monitoring Inference Aggregator,Detection Offset,Crop Visualization,Roboflow Dataset Upload,Dynamic Zone
Input and Output Bindings¶
The available connections depend on its binding kinds. Check what binding kinds
SAM 3 in version v1 has.
Bindings
-
input
images(image): The image to infer on..model_id(roboflow_model_id): model version. You only need to change this for fine tuned sam3 models..class_names(Union[string,list_of_values]): List of classes to recognise.threshold(float): Threshold for predicted mask scores.
-
output
predictions(instance_segmentation_prediction): Prediction with detected bounding boxes and segmentation masks in form of sv.Detections(...) object.
Example JSON definition of step SAM 3 in version v1
{
"name": "<your_step_name_here>",
"type": "roboflow_core/sam3@v1",
"images": "$inputs.image",
"model_id": "sam3/sam3_final",
"class_names": [
"car",
"person"
],
"threshold": 0.3
}