SAM 3¶
v3¶
Class: SegmentAnything3BlockV3 (there are multiple versions of this block)
Source: inference.core.workflows.core_steps.models.foundation.segment_anything3.v3.SegmentAnything3BlockV3
Warning: This block has multiple versions. Please refer to the specific version for details. You can learn more about how versions work here: Versioning
Run Segment Anything 3 (SAM3), a zero-shot instance segmentation model, on an image.
You can use text prompts for open-vocabulary segmentation - just specify class names and SAM3 will segment those objects in the image.
This block supports two output formats: - rle (default): Returns masks in RLE (Run-Length Encoding) format, which is more memory-efficient - polygons: Returns polygon coordinates for each mask
RLE format is recommended for high-resolution images or workflows with many detections.
Type identifier¶
Use the following identifier in step "type" field: roboflow_core/sam3@v3to add the block as
as step in your workflow.
Properties¶
| Name | Type | Description | Refs |
|---|---|---|---|
name |
str |
Enter a unique identifier for this step.. | ❌ |
model_id |
str |
model version. You only need to change this for fine tuned sam3 models.. | ✅ |
class_names |
Optional[List[str], str] |
List of classes to recognise. | ✅ |
confidence |
float |
Minimum confidence threshold for predicted masks. | ✅ |
per_class_confidence |
List[float] |
List of confidence thresholds per class (must match class_names length). | ✅ |
apply_nms |
bool |
Whether to apply Non-Maximum Suppression across prompts. | ✅ |
nms_iou_threshold |
float |
IoU threshold for cross-prompt NMS. Must be in [0.0, 1.0]. | ✅ |
output_format |
str |
'rle' returns efficient RLE encoding (recommended), 'polygons' returns polygon coordinates. | ❌ |
The Refs column marks possibility to parametrise the property with dynamic values available
in workflow runtime. See Bindings for more info.
Available Connections¶
Compatible Blocks
Check what blocks you can connect to SAM 3 in version v3.
- inputs:
Anthropic Claude,Mask Visualization,Classification Label Visualization,Instance Segmentation Model,Detections Consensus,Webhook Sink,Multi-Label Classification Model,Dynamic Zone,Email Notification,QR Code Generator,Dynamic Crop,VLM As Detector,VLM As Detector,Google Gemini,Multi-Label Classification Model,LMM,Image Blur,Corner Visualization,Image Convert Grayscale,Stability AI Outpainting,Halo Visualization,Stability AI Inpainting,Object Detection Model,JSON Parser,Image Contours,Single-Label Classification Model,Trace Visualization,Google Vision OCR,Morphological Transformation,Triangle Visualization,Instance Segmentation Model,Clip Comparison,Relative Static Crop,CSV Formatter,Text Display,Stitch Images,Google Gemini,Camera Calibration,Grid Visualization,Local File Sink,Slack Notification,VLM As Classifier,Roboflow Dataset Upload,Camera Focus,PTZ Tracking (ONVIF).md),Color Visualization,Dot Visualization,Image Slicer,Polygon Visualization,Object Detection Model,Anthropic Claude,Llama 3.2 Vision,Line Counter Visualization,LMM For Classification,Keypoint Detection Model,Buffer,Contrast Equalization,Identify Changes,SIFT Comparison,Dimension Collapse,Camera Focus,Background Subtraction,Image Slicer,Circle Visualization,SIFT Comparison,Identify Outliers,Halo Visualization,Florence-2 Model,Blur Visualization,Label Visualization,Twilio SMS/MMS Notification,Clip Comparison,Email Notification,Ellipse Visualization,OpenAI,SIFT,Image Preprocessing,Model Monitoring Inference Aggregator,Single-Label Classification Model,Detections List Roll-Up,OpenAI,Image Threshold,Background Color Visualization,Model Comparison Visualization,Depth Estimation,Size Measurement,OpenAI,Motion Detection,Cosine Similarity,Keypoint Detection Model,CogVLM,Absolute Static Crop,Roboflow Custom Metadata,Gaze Detection,EasyOCR,Stitch OCR Detections,Perspective Correction,Anthropic Claude,Pixelate Visualization,Stability AI Image Generation,Reference Path Visualization,Keypoint Visualization,Polygon Visualization,Twilio SMS Notification,VLM As Classifier,Bounding Box Visualization,Polygon Zone Visualization,OCR Model,Icon Visualization,Crop Visualization,Stitch OCR Detections,Google Gemini,OpenAI,Florence-2 Model,Roboflow Dataset Upload - outputs:
Mask Visualization,Circle Visualization,Detections Consensus,Detections Merge,Halo Visualization,Dynamic Zone,Florence-2 Model,Blur Visualization,Dynamic Crop,Florence-2 Model,Label Visualization,Path Deviation,Detection Offset,Corner Visualization,Ellipse Visualization,Byte Tracker,Byte Tracker,Line Counter,Model Monitoring Inference Aggregator,Segment Anything 2 Model,Halo Visualization,Stability AI Inpainting,Detections List Roll-Up,Model Comparison Visualization,Background Color Visualization,Path Deviation,Trace Visualization,Size Measurement,Time in Zone,Line Counter,Triangle Visualization,Bounding Rectangle,Detections Stitch,Detections Filter,Roboflow Custom Metadata,Detections Stabilizer,Roboflow Dataset Upload,PTZ Tracking (ONVIF).md),Camera Focus,Perspective Correction,Color Visualization,Dot Visualization,Detections Combine,Polygon Visualization,Pixelate Visualization,Polygon Visualization,Bounding Box Visualization,Byte Tracker,Detection Event Log,Distance Measurement,Detections Classes Replacement,Overlap Filter,Icon Visualization,Crop Visualization,Time in Zone,Time in Zone,Velocity,Detections Transformation,Roboflow Dataset Upload
Input and Output Bindings¶
The available connections depend on its binding kinds. Check what binding kinds
SAM 3 in version v3 has.
Bindings
-
input
images(image): The image to infer on..model_id(roboflow_model_id): model version. You only need to change this for fine tuned sam3 models..class_names(Union[list_of_values,string]): List of classes to recognise.confidence(float): Minimum confidence threshold for predicted masks.per_class_confidence(list_of_values): List of confidence thresholds per class (must match class_names length).apply_nms(boolean): Whether to apply Non-Maximum Suppression across prompts.nms_iou_threshold(float): IoU threshold for cross-prompt NMS. Must be in [0.0, 1.0].
-
output
predictions(Union[rle_instance_segmentation_prediction,instance_segmentation_prediction]): Prediction with detected bounding boxes and RLE-encoded segmentation masks in form of sv.Detections(...) object ifrle_instance_segmentation_predictionor Prediction with detected bounding boxes and segmentation masks in form of sv.Detections(...) object ifinstance_segmentation_prediction.
Example JSON definition of step SAM 3 in version v3
{
"name": "<your_step_name_here>",
"type": "roboflow_core/sam3@v3",
"images": "$inputs.image",
"model_id": "sam3/sam3_final",
"class_names": [
"car",
"person"
],
"confidence": 0.3,
"per_class_confidence": [
0.3,
0.5,
0.7
],
"apply_nms": "<block_does_not_provide_example>",
"nms_iou_threshold": 0.5,
"output_format": "rle"
}
v2¶
Class: SegmentAnything3BlockV2 (there are multiple versions of this block)
Source: inference.core.workflows.core_steps.models.foundation.segment_anything3.v2.SegmentAnything3BlockV2
Warning: This block has multiple versions. Please refer to the specific version for details. You can learn more about how versions work here: Versioning
Run Segment Anything 3, a zero-shot instance segmentation model, on an image.
You can pass in boxes/predictions from other models as prompts, or use a text prompt for open-vocabulary segmentation. If you pass in box detections from another model, the class names of the boxes will be forwarded to the predicted masks.
Type identifier¶
Use the following identifier in step "type" field: roboflow_core/sam3@v2to add the block as
as step in your workflow.
Properties¶
| Name | Type | Description | Refs |
|---|---|---|---|
name |
str |
Enter a unique identifier for this step.. | ❌ |
model_id |
str |
model version. You only need to change this for fine tuned sam3 models.. | ✅ |
class_names |
Optional[List[str], str] |
List of classes to recognise. | ✅ |
confidence |
float |
Minimum confidence threshold for predicted masks. | ✅ |
per_class_confidence |
List[float] |
List of confidence thresholds per class (must match class_names length). | ✅ |
apply_nms |
bool |
Whether to apply Non-Maximum Suppression across prompts. | ✅ |
nms_iou_threshold |
float |
IoU threshold for cross-prompt NMS. Must be in [0.0, 1.0]. | ✅ |
The Refs column marks possibility to parametrise the property with dynamic values available
in workflow runtime. See Bindings for more info.
Available Connections¶
Compatible Blocks
Check what blocks you can connect to SAM 3 in version v2.
- inputs:
Anthropic Claude,Mask Visualization,Classification Label Visualization,Instance Segmentation Model,Detections Consensus,Webhook Sink,Multi-Label Classification Model,Dynamic Zone,Email Notification,QR Code Generator,Dynamic Crop,VLM As Detector,VLM As Detector,Google Gemini,Multi-Label Classification Model,LMM,Image Blur,Corner Visualization,Image Convert Grayscale,Stability AI Outpainting,Halo Visualization,Stability AI Inpainting,Object Detection Model,JSON Parser,Image Contours,Single-Label Classification Model,Trace Visualization,Google Vision OCR,Morphological Transformation,Triangle Visualization,Instance Segmentation Model,Clip Comparison,Relative Static Crop,CSV Formatter,Text Display,Stitch Images,Google Gemini,Camera Calibration,Grid Visualization,Local File Sink,Slack Notification,VLM As Classifier,Roboflow Dataset Upload,Camera Focus,PTZ Tracking (ONVIF).md),Color Visualization,Dot Visualization,Image Slicer,Polygon Visualization,Object Detection Model,Anthropic Claude,Llama 3.2 Vision,Line Counter Visualization,LMM For Classification,Keypoint Detection Model,Buffer,Contrast Equalization,Identify Changes,SIFT Comparison,Dimension Collapse,Camera Focus,Background Subtraction,Image Slicer,Circle Visualization,SIFT Comparison,Identify Outliers,Halo Visualization,Florence-2 Model,Blur Visualization,Label Visualization,Twilio SMS/MMS Notification,Clip Comparison,Email Notification,Ellipse Visualization,OpenAI,SIFT,Image Preprocessing,Model Monitoring Inference Aggregator,Single-Label Classification Model,Detections List Roll-Up,OpenAI,Image Threshold,Background Color Visualization,Model Comparison Visualization,Depth Estimation,Size Measurement,OpenAI,Motion Detection,Cosine Similarity,Keypoint Detection Model,CogVLM,Absolute Static Crop,Roboflow Custom Metadata,Gaze Detection,EasyOCR,Stitch OCR Detections,Perspective Correction,Anthropic Claude,Pixelate Visualization,Stability AI Image Generation,Reference Path Visualization,Keypoint Visualization,Polygon Visualization,Twilio SMS Notification,VLM As Classifier,Bounding Box Visualization,Polygon Zone Visualization,OCR Model,Icon Visualization,Crop Visualization,Stitch OCR Detections,Google Gemini,OpenAI,Florence-2 Model,Roboflow Dataset Upload - outputs:
Mask Visualization,Circle Visualization,Detections Consensus,Detections Merge,Halo Visualization,Dynamic Zone,Florence-2 Model,Blur Visualization,Florence-2 Model,Dynamic Crop,Label Visualization,Path Deviation,Detection Offset,Corner Visualization,Ellipse Visualization,Byte Tracker,Byte Tracker,Line Counter,Model Monitoring Inference Aggregator,Segment Anything 2 Model,Halo Visualization,Stability AI Inpainting,Detections List Roll-Up,Model Comparison Visualization,Background Color Visualization,Size Measurement,Trace Visualization,Path Deviation,Time in Zone,Line Counter,Triangle Visualization,Bounding Rectangle,Detections Stitch,Detections Filter,Roboflow Custom Metadata,Detections Stabilizer,Roboflow Dataset Upload,PTZ Tracking (ONVIF).md),Camera Focus,Perspective Correction,Color Visualization,Detections Combine,Pixelate Visualization,Polygon Visualization,Dot Visualization,Polygon Visualization,Bounding Box Visualization,Detection Event Log,Byte Tracker,Distance Measurement,Detections Classes Replacement,Overlap Filter,Icon Visualization,Crop Visualization,Time in Zone,Time in Zone,Velocity,Detections Transformation,Roboflow Dataset Upload
Input and Output Bindings¶
The available connections depend on its binding kinds. Check what binding kinds
SAM 3 in version v2 has.
Bindings
-
input
images(image): The image to infer on..model_id(roboflow_model_id): model version. You only need to change this for fine tuned sam3 models..class_names(Union[list_of_values,string]): List of classes to recognise.confidence(float): Minimum confidence threshold for predicted masks.per_class_confidence(list_of_values): List of confidence thresholds per class (must match class_names length).apply_nms(boolean): Whether to apply Non-Maximum Suppression across prompts.nms_iou_threshold(float): IoU threshold for cross-prompt NMS. Must be in [0.0, 1.0].
-
output
predictions(instance_segmentation_prediction): Prediction with detected bounding boxes and segmentation masks in form of sv.Detections(...) object.
Example JSON definition of step SAM 3 in version v2
{
"name": "<your_step_name_here>",
"type": "roboflow_core/sam3@v2",
"images": "$inputs.image",
"model_id": "sam3/sam3_final",
"class_names": [
"car",
"person"
],
"confidence": 0.3,
"per_class_confidence": [
0.3,
0.5,
0.7
],
"apply_nms": "<block_does_not_provide_example>",
"nms_iou_threshold": 0.5
}
v1¶
Class: SegmentAnything3BlockV1 (there are multiple versions of this block)
Source: inference.core.workflows.core_steps.models.foundation.segment_anything3.v1.SegmentAnything3BlockV1
Warning: This block has multiple versions. Please refer to the specific version for details. You can learn more about how versions work here: Versioning
Run Segment Anything 3, a zero-shot instance segmentation model, on an image.
You can pass in boxes/predictions from other models as prompts, or use a text prompt for open-vocabulary segmentation. If you pass in box detections from another model, the class names of the boxes will be forwarded to the predicted masks.
Type identifier¶
Use the following identifier in step "type" field: roboflow_core/sam3@v1to add the block as
as step in your workflow.
Properties¶
| Name | Type | Description | Refs |
|---|---|---|---|
name |
str |
Enter a unique identifier for this step.. | ❌ |
model_id |
str |
model version. You only need to change this for fine tuned sam3 models.. | ✅ |
class_names |
Optional[List[str], str] |
List of classes to recognise. | ✅ |
threshold |
float |
Threshold for predicted mask scores. | ✅ |
The Refs column marks possibility to parametrise the property with dynamic values available
in workflow runtime. See Bindings for more info.
Available Connections¶
Compatible Blocks
Check what blocks you can connect to SAM 3 in version v1.
- inputs:
Anthropic Claude,Mask Visualization,Classification Label Visualization,Instance Segmentation Model,Webhook Sink,Multi-Label Classification Model,Dynamic Zone,Email Notification,QR Code Generator,Dynamic Crop,VLM As Detector,Google Gemini,Multi-Label Classification Model,LMM,Image Blur,Corner Visualization,Image Convert Grayscale,Stability AI Outpainting,Halo Visualization,Stability AI Inpainting,Object Detection Model,Image Contours,Single-Label Classification Model,Trace Visualization,Google Vision OCR,Morphological Transformation,Triangle Visualization,Instance Segmentation Model,Clip Comparison,Relative Static Crop,CSV Formatter,Text Display,Stitch Images,Google Gemini,Camera Calibration,Grid Visualization,Local File Sink,Slack Notification,VLM As Classifier,Roboflow Dataset Upload,Camera Focus,Color Visualization,Dot Visualization,Image Slicer,Polygon Visualization,Object Detection Model,Anthropic Claude,Llama 3.2 Vision,Line Counter Visualization,LMM For Classification,Keypoint Detection Model,Buffer,Contrast Equalization,Identify Changes,SIFT Comparison,Dimension Collapse,Camera Focus,Background Subtraction,Image Slicer,Circle Visualization,Halo Visualization,Florence-2 Model,Blur Visualization,Label Visualization,Twilio SMS/MMS Notification,Clip Comparison,Email Notification,Ellipse Visualization,OpenAI,SIFT,Image Preprocessing,Model Monitoring Inference Aggregator,Single-Label Classification Model,Detections List Roll-Up,OpenAI,Image Threshold,Background Color Visualization,Model Comparison Visualization,Depth Estimation,Size Measurement,OpenAI,Motion Detection,Cosine Similarity,Keypoint Detection Model,CogVLM,Absolute Static Crop,Roboflow Custom Metadata,Gaze Detection,EasyOCR,Stitch OCR Detections,Perspective Correction,Anthropic Claude,Pixelate Visualization,Stability AI Image Generation,Reference Path Visualization,Keypoint Visualization,Polygon Visualization,Twilio SMS Notification,Bounding Box Visualization,Polygon Zone Visualization,OCR Model,Icon Visualization,Crop Visualization,Stitch OCR Detections,Google Gemini,OpenAI,Florence-2 Model,Roboflow Dataset Upload - outputs:
Mask Visualization,Circle Visualization,Detections Consensus,Detections Merge,Halo Visualization,Dynamic Zone,Florence-2 Model,Blur Visualization,Florence-2 Model,Dynamic Crop,Label Visualization,Path Deviation,Detection Offset,Corner Visualization,Ellipse Visualization,Byte Tracker,Byte Tracker,Line Counter,Model Monitoring Inference Aggregator,Segment Anything 2 Model,Halo Visualization,Stability AI Inpainting,Detections List Roll-Up,Model Comparison Visualization,Background Color Visualization,Size Measurement,Trace Visualization,Path Deviation,Time in Zone,Line Counter,Triangle Visualization,Bounding Rectangle,Detections Stitch,Detections Filter,Roboflow Custom Metadata,Detections Stabilizer,Roboflow Dataset Upload,PTZ Tracking (ONVIF).md),Camera Focus,Perspective Correction,Color Visualization,Detections Combine,Pixelate Visualization,Polygon Visualization,Dot Visualization,Polygon Visualization,Bounding Box Visualization,Detection Event Log,Byte Tracker,Distance Measurement,Detections Classes Replacement,Overlap Filter,Icon Visualization,Crop Visualization,Time in Zone,Time in Zone,Velocity,Detections Transformation,Roboflow Dataset Upload
Input and Output Bindings¶
The available connections depend on its binding kinds. Check what binding kinds
SAM 3 in version v1 has.
Bindings
-
input
images(image): The image to infer on..model_id(roboflow_model_id): model version. You only need to change this for fine tuned sam3 models..class_names(Union[list_of_values,string]): List of classes to recognise.threshold(float): Threshold for predicted mask scores.
-
output
predictions(instance_segmentation_prediction): Prediction with detected bounding boxes and segmentation masks in form of sv.Detections(...) object.
Example JSON definition of step SAM 3 in version v1
{
"name": "<your_step_name_here>",
"type": "roboflow_core/sam3@v1",
"images": "$inputs.image",
"model_id": "sam3/sam3_final",
"class_names": [
"car",
"person"
],
"threshold": 0.3
}