YOLO-World Model¶
Class: YoloWorldModelBlockV1
Source: inference.core.workflows.core_steps.models.foundation.yolo_world.v1.YoloWorldModelBlockV1
Run YOLO-World, a zero-shot object detection model, on an image.
YOLO-World accepts one or more text classes you want to identify in an image. The model returns the location of objects that meet the specified class, if YOLO-World is able to identify objects of that class.
We recommend experimenting with YOLO-World to evaluate the model on your use case before using this block in production. For example on how to effectively prompt YOLO-World, refer to the Roboflow YOLO-World prompting guide.
Type identifier¶
Use the following identifier in step "type" field: roboflow_core/yolo_world_model@v1to add the block as
as step in your workflow.
Properties¶
| Name | Type | Description | Refs |
|---|---|---|---|
name |
str |
Enter a unique identifier for this step.. | ❌ |
class_names |
List[str] |
One or more classes that you want YOLO-World to detect. The model accepts any string as an input, though does best with short descriptions of common objects.. | ✅ |
version |
str |
Variant of YoloWorld model. | ✅ |
confidence |
float |
Confidence threshold for detections. | ✅ |
The Refs column marks possibility to parametrise the property with dynamic values available
in workflow runtime. See Bindings for more info.
Available Connections¶
Compatible Blocks
Check what blocks you can connect to YOLO-World Model in version v1.
- inputs:
Roboflow Dataset Upload,Line Counter Visualization,Stability AI Outpainting,Google Gemma API,Object Detection Model,Email Notification,Image Slicer,OCR Model,Google Vision OCR,Identify Outliers,Image Preprocessing,Google Gemini,Instance Segmentation Model,EasyOCR,Color Visualization,OpenAI,Ellipse Visualization,Polygon Visualization,Anthropic Claude,Relative Static Crop,Detections Consensus,Webhook Sink,Model Comparison Visualization,Trace Visualization,Stitch OCR Detections,Camera Focus,Qwen 3.5 API,Roboflow Custom Metadata,OpenAI,Buffer,Single-Label Classification Model,Detections List Roll-Up,Size Measurement,Image Threshold,VLM As Classifier,Stitch Images,Heatmap Visualization,Qwen 3.6 API,SIFT Comparison,Morphological Transformation,Florence-2 Model,Halo Visualization,CogVLM,Crop Visualization,Camera Calibration,Florence-2 Model,GLM-OCR,Dot Visualization,S3 Sink,Twilio SMS Notification,Icon Visualization,Model Monitoring Inference Aggregator,Google Gemini,Local File Sink,Roboflow Dataset Upload,Dynamic Zone,Clip Comparison,Image Contours,Pixelate Visualization,Twilio SMS/MMS Notification,Polygon Zone Visualization,Reference Path Visualization,Dimension Collapse,Motion Detection,Blur Visualization,Anthropic Claude,Background Subtraction,Text Display,Clip Comparison,CSV Formatter,VLM As Detector,LMM,Stability AI Image Generation,Perspective Correction,Anthropic Claude,Bounding Box Visualization,Depth Estimation,Identify Changes,Classification Label Visualization,Image Slicer,Absolute Static Crop,Image Blur,Stability AI Inpainting,Multi-Label Classification Model,Polygon Visualization,Image Convert Grayscale,SIFT,Roboflow Vision Events,OpenAI,Google Gemini,Label Visualization,Corner Visualization,Grid Visualization,Dynamic Crop,Contrast Equalization,Keypoint Visualization,Triangle Visualization,Qwen3.5-VL,QR Code Generator,Halo Visualization,Circle Visualization,Camera Focus,Mask Visualization,LMM For Classification,Morphological Transformation,OpenAI,Contrast Enhancement,MoonshotAI Kimi,Keypoint Detection Model,Llama 3.2 Vision,Background Color Visualization,Email Notification,Slack Notification,Stitch OCR Detections - outputs:
Detections Stabilizer,Detections Stitch,Model Monitoring Inference Aggregator,Detections Filter,Roboflow Dataset Upload,Roboflow Dataset Upload,Icon Visualization,Pixelate Visualization,Line Counter,Time in Zone,Distance Measurement,Color Visualization,Detections Combine,Blur Visualization,SAM2 Video Tracker,Detection Event Log,Ellipse Visualization,ByteTrack Tracker,Byte Tracker,Byte Tracker,Time in Zone,Detections Merge,Detections Classes Replacement,Detections Consensus,Perspective Correction,Overlap Filter,Model Comparison Visualization,Velocity,Stitch OCR Detections,Trace Visualization,Camera Focus,Roboflow Custom Metadata,Byte Tracker,Bounding Box Visualization,Line Counter,Detection Offset,Roboflow Vision Events,Detections List Roll-Up,Size Measurement,Label Visualization,Corner Visualization,Mask Area Measurement,Heatmap Visualization,Dynamic Crop,Per-Class Confidence Filter,SORT Tracker,Triangle Visualization,Circle Visualization,Florence-2 Model,Segment Anything 2 Model,Detections Transformation,Crop Visualization,Florence-2 Model,Path Deviation,Background Color Visualization,Time in Zone,Dot Visualization,OC-SORT Tracker,PTZ Tracking (ONVIF),Path Deviation,Stitch OCR Detections
Input and Output Bindings¶
The available connections depend on its binding kinds. Check what binding kinds
YOLO-World Model in version v1 has.
Bindings
-
input
images(image): The image to infer on..class_names(list_of_values): One or more classes that you want YOLO-World to detect. The model accepts any string as an input, though does best with short descriptions of common objects..version(string): Variant of YoloWorld model.confidence(float_zero_to_one): Confidence threshold for detections.
-
output
predictions(object_detection_prediction): Prediction with detected bounding boxes in form of sv.Detections(...) object.
Example JSON definition of step YOLO-World Model in version v1
{
"name": "<your_step_name_here>",
"type": "roboflow_core/yolo_world_model@v1",
"images": "$inputs.image",
"class_names": [
"person",
"car",
"license plate"
],
"version": "v2-s",
"confidence": 0.005
}