SAM 3¶
v3¶
Class: SegmentAnything3BlockV3 (there are multiple versions of this block)
Source: inference.core.workflows.core_steps.models.foundation.segment_anything3.v3.SegmentAnything3BlockV3
Warning: This block has multiple versions. Please refer to the specific version for details. You can learn more about how versions work here: Versioning
Run Segment Anything 3 (SAM3), a zero-shot instance segmentation model, on an image.
You can use text prompts for open-vocabulary segmentation - just specify class names and SAM3 will segment those objects in the image.
This block supports two output formats: - rle (default): Returns masks in RLE (Run-Length Encoding) format, which is more memory-efficient - polygons: Returns polygon coordinates for each mask
RLE format is recommended for high-resolution images or workflows with many detections.
Type identifier¶
Use the following identifier in step "type" field: roboflow_core/sam3@v3to add the block as
as step in your workflow.
Properties¶
| Name | Type | Description | Refs |
|---|---|---|---|
name |
str |
Enter a unique identifier for this step.. | ❌ |
model_id |
str |
model version. You only need to change this for fine tuned sam3 models.. | ✅ |
class_names |
Optional[List[str], str] |
List of classes to recognise. | ✅ |
confidence |
float |
Minimum confidence threshold for predicted masks. | ✅ |
per_class_confidence |
List[float] |
List of confidence thresholds per class (must match class_names length). | ✅ |
apply_nms |
bool |
Whether to apply Non-Maximum Suppression across prompts. | ✅ |
nms_iou_threshold |
float |
IoU threshold for cross-prompt NMS. Must be in [0.0, 1.0]. | ✅ |
output_format |
str |
'rle' returns efficient RLE encoding (recommended), 'polygons' returns polygon coordinates. | ❌ |
The Refs column marks possibility to parametrise the property with dynamic values available
in workflow runtime. See Bindings for more info.
Available Connections¶
Compatible Blocks
Check what blocks you can connect to SAM 3 in version v3.
- inputs:
Detections Consensus,Llama 3.2 Vision,Cosine Similarity,Blur Visualization,Dimension Collapse,Perspective Correction,Polygon Zone Visualization,Bounding Box Visualization,QR Code Generator,Pixelate Visualization,Trace Visualization,Roboflow Custom Metadata,Image Threshold,Polygon Visualization,Dynamic Crop,Icon Visualization,Image Slicer,Identify Outliers,Stability AI Outpainting,Model Comparison Visualization,Single-Label Classification Model,Dynamic Zone,Clip Comparison,LMM,OpenAI,Classification Label Visualization,Stitch Images,Florence-2 Model,Mask Visualization,Single-Label Classification Model,Size Measurement,Relative Static Crop,Absolute Static Crop,SIFT Comparison,Google Gemini,Circle Visualization,Florence-2 Model,LMM For Classification,Ellipse Visualization,Image Convert Grayscale,Object Detection Model,OCR Model,Image Preprocessing,Color Visualization,Image Blur,Stability AI Image Generation,Anthropic Claude,Google Vision OCR,Keypoint Visualization,Camera Calibration,Local File Sink,Keypoint Detection Model,EasyOCR,Image Slicer,Email Notification,VLM as Detector,VLM as Detector,Roboflow Dataset Upload,Gaze Detection,Background Color Visualization,Triangle Visualization,Slack Notification,Keypoint Detection Model,Multi-Label Classification Model,Object Detection Model,Halo Visualization,Corner Visualization,Google Gemini,Model Monitoring Inference Aggregator,Roboflow Dataset Upload,Dot Visualization,Image Contours,Multi-Label Classification Model,Twilio SMS Notification,Instance Segmentation Model,VLM as Classifier,CSV Formatter,Reference Path Visualization,Morphological Transformation,Motion Detection,OpenAI,Webhook Sink,PTZ Tracking (ONVIF).md),Instance Segmentation Model,Contrast Equalization,Camera Focus,Stitch OCR Detections,Stability AI Inpainting,JSON Parser,CogVLM,Clip Comparison,Line Counter Visualization,Identify Changes,Email Notification,Crop Visualization,Grid Visualization,VLM as Classifier,OpenAI,Buffer,SIFT,Depth Estimation,Background Subtraction,Label Visualization,Anthropic Claude,SIFT Comparison,OpenAI - outputs:
Line Counter,Detections Consensus,Path Deviation,Model Monitoring Inference Aggregator,Roboflow Dataset Upload,Blur Visualization,Dot Visualization,Detections Merge,Perspective Correction,Byte Tracker,Bounding Box Visualization,Pixelate Visualization,Distance Measurement,Trace Visualization,Roboflow Custom Metadata,Velocity,Byte Tracker,Detections Transformation,Segment Anything 2 Model,PTZ Tracking (ONVIF).md),Polygon Visualization,Icon Visualization,Dynamic Crop,Detections Classes Replacement,Model Comparison Visualization,Dynamic Zone,Detections Stitch,Size Measurement,Mask Visualization,Florence-2 Model,Stability AI Inpainting,Time in Zone,Circle Visualization,Florence-2 Model,Ellipse Visualization,Time in Zone,Path Deviation,Crop Visualization,Color Visualization,Bounding Rectangle,Line Counter,Detections Combine,Label Visualization,Byte Tracker,Roboflow Dataset Upload,Overlap Filter,Triangle Visualization,Background Color Visualization,Detections Filter,Time in Zone,Detection Offset,Halo Visualization,Detections Stabilizer,Corner Visualization
Input and Output Bindings¶
The available connections depend on its binding kinds. Check what binding kinds
SAM 3 in version v3 has.
Bindings
-
input
images(image): The image to infer on..model_id(roboflow_model_id): model version. You only need to change this for fine tuned sam3 models..class_names(Union[list_of_values,string]): List of classes to recognise.confidence(float): Minimum confidence threshold for predicted masks.per_class_confidence(list_of_values): List of confidence thresholds per class (must match class_names length).apply_nms(boolean): Whether to apply Non-Maximum Suppression across prompts.nms_iou_threshold(float): IoU threshold for cross-prompt NMS. Must be in [0.0, 1.0].
-
output
predictions(Union[rle_instance_segmentation_prediction,instance_segmentation_prediction]): Prediction with detected bounding boxes and RLE-encoded segmentation masks in form of sv.Detections(...) object ifrle_instance_segmentation_predictionor Prediction with detected bounding boxes and segmentation masks in form of sv.Detections(...) object ifinstance_segmentation_prediction.
Example JSON definition of step SAM 3 in version v3
{
"name": "<your_step_name_here>",
"type": "roboflow_core/sam3@v3",
"images": "$inputs.image",
"model_id": "sam3/sam3_final",
"class_names": [
"car",
"person"
],
"confidence": 0.3,
"per_class_confidence": [
0.3,
0.5,
0.7
],
"apply_nms": "<block_does_not_provide_example>",
"nms_iou_threshold": 0.5,
"output_format": "rle"
}
v2¶
Class: SegmentAnything3BlockV2 (there are multiple versions of this block)
Source: inference.core.workflows.core_steps.models.foundation.segment_anything3.v2.SegmentAnything3BlockV2
Warning: This block has multiple versions. Please refer to the specific version for details. You can learn more about how versions work here: Versioning
Run Segment Anything 3, a zero-shot instance segmentation model, on an image.
You can pass in boxes/predictions from other models as prompts, or use a text prompt for open-vocabulary segmentation. If you pass in box detections from another model, the class names of the boxes will be forwarded to the predicted masks.
Type identifier¶
Use the following identifier in step "type" field: roboflow_core/sam3@v2to add the block as
as step in your workflow.
Properties¶
| Name | Type | Description | Refs |
|---|---|---|---|
name |
str |
Enter a unique identifier for this step.. | ❌ |
model_id |
str |
model version. You only need to change this for fine tuned sam3 models.. | ✅ |
class_names |
Optional[List[str], str] |
List of classes to recognise. | ✅ |
confidence |
float |
Minimum confidence threshold for predicted masks. | ✅ |
per_class_confidence |
List[float] |
List of confidence thresholds per class (must match class_names length). | ✅ |
apply_nms |
bool |
Whether to apply Non-Maximum Suppression across prompts. | ✅ |
nms_iou_threshold |
float |
IoU threshold for cross-prompt NMS. Must be in [0.0, 1.0]. | ✅ |
The Refs column marks possibility to parametrise the property with dynamic values available
in workflow runtime. See Bindings for more info.
Available Connections¶
Compatible Blocks
Check what blocks you can connect to SAM 3 in version v2.
- inputs:
Detections Consensus,Llama 3.2 Vision,Cosine Similarity,Blur Visualization,Dimension Collapse,Perspective Correction,Polygon Zone Visualization,Bounding Box Visualization,QR Code Generator,Pixelate Visualization,Trace Visualization,Roboflow Custom Metadata,Image Threshold,Polygon Visualization,Dynamic Crop,Icon Visualization,Image Slicer,Identify Outliers,Stability AI Outpainting,Model Comparison Visualization,Single-Label Classification Model,Dynamic Zone,Clip Comparison,LMM,OpenAI,Classification Label Visualization,Stitch Images,Florence-2 Model,Mask Visualization,Single-Label Classification Model,Size Measurement,Relative Static Crop,Absolute Static Crop,SIFT Comparison,Google Gemini,Circle Visualization,Florence-2 Model,LMM For Classification,Ellipse Visualization,Image Convert Grayscale,Object Detection Model,OCR Model,Image Preprocessing,Color Visualization,Image Blur,Stability AI Image Generation,Anthropic Claude,Google Vision OCR,Keypoint Visualization,Camera Calibration,Local File Sink,Keypoint Detection Model,EasyOCR,Image Slicer,Email Notification,VLM as Detector,VLM as Detector,Roboflow Dataset Upload,Gaze Detection,Background Color Visualization,Triangle Visualization,Slack Notification,Keypoint Detection Model,Multi-Label Classification Model,Object Detection Model,Halo Visualization,Corner Visualization,Google Gemini,Model Monitoring Inference Aggregator,Roboflow Dataset Upload,Dot Visualization,Image Contours,Multi-Label Classification Model,Twilio SMS Notification,Instance Segmentation Model,VLM as Classifier,CSV Formatter,Reference Path Visualization,Morphological Transformation,Motion Detection,OpenAI,Webhook Sink,PTZ Tracking (ONVIF).md),Instance Segmentation Model,Contrast Equalization,Camera Focus,Stitch OCR Detections,Stability AI Inpainting,JSON Parser,CogVLM,Clip Comparison,Line Counter Visualization,Identify Changes,Email Notification,Crop Visualization,Grid Visualization,VLM as Classifier,OpenAI,Buffer,SIFT,Depth Estimation,Background Subtraction,Label Visualization,Anthropic Claude,SIFT Comparison,OpenAI - outputs:
Line Counter,Detections Consensus,Path Deviation,Model Monitoring Inference Aggregator,Roboflow Dataset Upload,Blur Visualization,Dot Visualization,Detections Merge,Perspective Correction,Byte Tracker,Bounding Box Visualization,Distance Measurement,Pixelate Visualization,Trace Visualization,Roboflow Custom Metadata,Velocity,Byte Tracker,Detections Transformation,PTZ Tracking (ONVIF).md),Segment Anything 2 Model,Polygon Visualization,Dynamic Crop,Icon Visualization,Detections Classes Replacement,Model Comparison Visualization,Dynamic Zone,Detections Stitch,Size Measurement,Mask Visualization,Florence-2 Model,Stability AI Inpainting,Time in Zone,Circle Visualization,Florence-2 Model,Path Deviation,Ellipse Visualization,Time in Zone,Crop Visualization,Color Visualization,Bounding Rectangle,Line Counter,Label Visualization,Detections Combine,Byte Tracker,Roboflow Dataset Upload,Overlap Filter,Triangle Visualization,Background Color Visualization,Detections Filter,Detection Offset,Time in Zone,Halo Visualization,Detections Stabilizer,Corner Visualization
Input and Output Bindings¶
The available connections depend on its binding kinds. Check what binding kinds
SAM 3 in version v2 has.
Bindings
-
input
images(image): The image to infer on..model_id(roboflow_model_id): model version. You only need to change this for fine tuned sam3 models..class_names(Union[list_of_values,string]): List of classes to recognise.confidence(float): Minimum confidence threshold for predicted masks.per_class_confidence(list_of_values): List of confidence thresholds per class (must match class_names length).apply_nms(boolean): Whether to apply Non-Maximum Suppression across prompts.nms_iou_threshold(float): IoU threshold for cross-prompt NMS. Must be in [0.0, 1.0].
-
output
predictions(instance_segmentation_prediction): Prediction with detected bounding boxes and segmentation masks in form of sv.Detections(...) object.
Example JSON definition of step SAM 3 in version v2
{
"name": "<your_step_name_here>",
"type": "roboflow_core/sam3@v2",
"images": "$inputs.image",
"model_id": "sam3/sam3_final",
"class_names": [
"car",
"person"
],
"confidence": 0.3,
"per_class_confidence": [
0.3,
0.5,
0.7
],
"apply_nms": "<block_does_not_provide_example>",
"nms_iou_threshold": 0.5
}
v1¶
Class: SegmentAnything3BlockV1 (there are multiple versions of this block)
Source: inference.core.workflows.core_steps.models.foundation.segment_anything3.v1.SegmentAnything3BlockV1
Warning: This block has multiple versions. Please refer to the specific version for details. You can learn more about how versions work here: Versioning
Run Segment Anything 3, a zero-shot instance segmentation model, on an image.
You can pass in boxes/predictions from other models as prompts, or use a text prompt for open-vocabulary segmentation. If you pass in box detections from another model, the class names of the boxes will be forwarded to the predicted masks.
Type identifier¶
Use the following identifier in step "type" field: roboflow_core/sam3@v1to add the block as
as step in your workflow.
Properties¶
| Name | Type | Description | Refs |
|---|---|---|---|
name |
str |
Enter a unique identifier for this step.. | ❌ |
model_id |
str |
model version. You only need to change this for fine tuned sam3 models.. | ✅ |
class_names |
Optional[List[str], str] |
List of classes to recognise. | ✅ |
threshold |
float |
Threshold for predicted mask scores. | ✅ |
The Refs column marks possibility to parametrise the property with dynamic values available
in workflow runtime. See Bindings for more info.
Available Connections¶
Compatible Blocks
Check what blocks you can connect to SAM 3 in version v1.
- inputs:
Llama 3.2 Vision,Cosine Similarity,Blur Visualization,Dimension Collapse,Perspective Correction,Polygon Zone Visualization,Bounding Box Visualization,QR Code Generator,Pixelate Visualization,Trace Visualization,Roboflow Custom Metadata,Image Threshold,Polygon Visualization,Dynamic Crop,Icon Visualization,Image Slicer,Stability AI Outpainting,Model Comparison Visualization,Single-Label Classification Model,Dynamic Zone,Clip Comparison,LMM,OpenAI,Classification Label Visualization,Stitch Images,Florence-2 Model,Mask Visualization,Single-Label Classification Model,Size Measurement,Relative Static Crop,Absolute Static Crop,SIFT Comparison,Google Gemini,Circle Visualization,Florence-2 Model,LMM For Classification,Ellipse Visualization,Image Convert Grayscale,Object Detection Model,OCR Model,Image Preprocessing,Color Visualization,Image Blur,Stability AI Image Generation,Anthropic Claude,Google Vision OCR,Keypoint Visualization,Camera Calibration,Local File Sink,Keypoint Detection Model,EasyOCR,Image Slicer,Email Notification,VLM as Detector,Roboflow Dataset Upload,Gaze Detection,Background Color Visualization,Triangle Visualization,Slack Notification,Keypoint Detection Model,Multi-Label Classification Model,Object Detection Model,Halo Visualization,Corner Visualization,Google Gemini,Model Monitoring Inference Aggregator,Roboflow Dataset Upload,Dot Visualization,Image Contours,Multi-Label Classification Model,Twilio SMS Notification,Instance Segmentation Model,VLM as Classifier,CSV Formatter,Reference Path Visualization,Morphological Transformation,Motion Detection,OpenAI,Webhook Sink,Instance Segmentation Model,Contrast Equalization,Camera Focus,Stitch OCR Detections,Stability AI Inpainting,CogVLM,Clip Comparison,Line Counter Visualization,Identify Changes,Email Notification,Crop Visualization,Grid Visualization,OpenAI,Buffer,SIFT,Depth Estimation,Background Subtraction,Label Visualization,Anthropic Claude,OpenAI - outputs:
Line Counter,Detections Consensus,Path Deviation,Model Monitoring Inference Aggregator,Roboflow Dataset Upload,Blur Visualization,Dot Visualization,Detections Merge,Perspective Correction,Byte Tracker,Bounding Box Visualization,Distance Measurement,Pixelate Visualization,Trace Visualization,Roboflow Custom Metadata,Velocity,Byte Tracker,Detections Transformation,PTZ Tracking (ONVIF).md),Segment Anything 2 Model,Polygon Visualization,Dynamic Crop,Icon Visualization,Detections Classes Replacement,Model Comparison Visualization,Dynamic Zone,Detections Stitch,Size Measurement,Mask Visualization,Florence-2 Model,Stability AI Inpainting,Time in Zone,Circle Visualization,Florence-2 Model,Path Deviation,Ellipse Visualization,Time in Zone,Crop Visualization,Color Visualization,Bounding Rectangle,Line Counter,Label Visualization,Detections Combine,Byte Tracker,Roboflow Dataset Upload,Overlap Filter,Triangle Visualization,Background Color Visualization,Detections Filter,Detection Offset,Time in Zone,Halo Visualization,Detections Stabilizer,Corner Visualization
Input and Output Bindings¶
The available connections depend on its binding kinds. Check what binding kinds
SAM 3 in version v1 has.
Bindings
-
input
images(image): The image to infer on..model_id(roboflow_model_id): model version. You only need to change this for fine tuned sam3 models..class_names(Union[list_of_values,string]): List of classes to recognise.threshold(float): Threshold for predicted mask scores.
-
output
predictions(instance_segmentation_prediction): Prediction with detected bounding boxes and segmentation masks in form of sv.Detections(...) object.
Example JSON definition of step SAM 3 in version v1
{
"name": "<your_step_name_here>",
"type": "roboflow_core/sam3@v1",
"images": "$inputs.image",
"model_id": "sam3/sam3_final",
"class_names": [
"car",
"person"
],
"threshold": 0.3
}