OCR Model¶
Class: OCRModelBlockV1
Source: inference.core.workflows.core_steps.models.foundation.ocr.v1.OCRModelBlockV1
Retrieve the characters in an image using Optical Character Recognition (OCR).
This block returns the text within an image.
You may want to use this block in combination with a detections-based block (i.e. ObjectDetectionBlock). An object detection model could isolate specific regions from an image (i.e. a shipping container ID in a logistics use case) for further processing. You can then use a DynamicCropBlock to crop the region of interest before running OCR.
Using a detections model then cropping detections allows you to isolate your analysis on particular regions of an image.
Type identifier¶
Use the following identifier in step "type"
field: roboflow_core/ocr_model@v1
to add the block as
as step in your workflow.
Properties¶
Name | Type | Description | Refs |
---|---|---|---|
name |
str |
Unique name of step in workflows. | ❌ |
The Refs column marks possibility to parametrise the property with dynamic values available
in workflow
runtime. See Bindings for more info.
Available Connections¶
Compatible Blocks
Check what blocks you can connect to OCR Model
in version v1
.
- inputs:
Image Threshold
,Image Convert Grayscale
,SIFT Comparison
,Absolute Static Crop
,Pixelate Visualization
,SIFT
,Line Counter Visualization
,Stitch Images
,Stability AI Image Generation
,Dot Visualization
,Mask Visualization
,Image Slicer
,Background Color Visualization
,Triangle Visualization
,Image Slicer
,Image Blur
,Camera Focus
,Grid Visualization
,Polygon Zone Visualization
,Blur Visualization
,Crop Visualization
,Label Visualization
,Classification Label Visualization
,Depth Estimation
,Image Preprocessing
,Model Comparison Visualization
,Stability AI Inpainting
,Reference Path Visualization
,Ellipse Visualization
,Bounding Box Visualization
,Halo Visualization
,Image Contours
,Corner Visualization
,Camera Calibration
,Circle Visualization
,Perspective Correction
,Dynamic Crop
,Polygon Visualization
,Relative Static Crop
,Trace Visualization
,Keypoint Visualization
,Color Visualization
- outputs:
CogVLM
,Anthropic Claude
,Cache Set
,Google Vision OCR
,OpenAI
,Detections Classes Replacement
,Florence-2 Model
,Distance Measurement
,OpenAI
,Stability AI Image Generation
,YOLO-World Model
,Mask Visualization
,Triangle Visualization
,Path Deviation
,Polygon Zone Visualization
,Model Comparison Visualization
,Crop Visualization
,LMM
,Classification Label Visualization
,Segment Anything 2 Model
,Reference Path Visualization
,Twilio SMS Notification
,Line Counter
,Google Gemini
,Bounding Box Visualization
,Roboflow Custom Metadata
,Size Measurement
,Circle Visualization
,Perspective Correction
,Slack Notification
,Pixel Color Count
,Polygon Visualization
,Instance Segmentation Model
,Trace Visualization
,Webhook Sink
,Color Visualization
,LMM For Classification
,Image Threshold
,Detections Stitch
,SIFT Comparison
,Line Counter Visualization
,Line Counter
,Dot Visualization
,Instance Segmentation Model
,Florence-2 Model
,Background Color Visualization
,Model Monitoring Inference Aggregator
,Roboflow Dataset Upload
,Roboflow Dataset Upload
,Image Blur
,Time in Zone
,Email Notification
,Path Deviation
,CLIP Embedding Model
,Stability AI Inpainting
,Label Visualization
,Llama 3.2 Vision
,Image Preprocessing
,Ellipse Visualization
,ONVIF Control
,Halo Visualization
,Corner Visualization
,Clip Comparison
,Cache Get
,Time in Zone
,Dynamic Crop
,OpenAI
,Keypoint Visualization
,Local File Sink
Input and Output Bindings¶
The available connections depend on its binding kinds. Check what binding kinds
OCR Model
in version v1
has.
Bindings
-
input
images
(image
): The image to infer on..
-
output
result
(string
): String value.parent_id
(parent_id
): Identifier of parent for step output.root_parent_id
(parent_id
): Identifier of parent for step output.prediction_type
(prediction_type
): String value with type of prediction.
Example JSON definition of step OCR Model
in version v1
{
"name": "<your_step_name_here>",
"type": "roboflow_core/ocr_model@v1",
"images": "$inputs.image"
}