OCR Model¶
Class: OCRModelBlockV1
Source: inference.core.workflows.core_steps.models.foundation.ocr.v1.OCRModelBlockV1
Retrieve the characters in an image using Optical Character Recognition (OCR).
This block returns the text within an image.
You may want to use this block in combination with a detections-based block (i.e. ObjectDetectionBlock). An object detection model could isolate specific regions from an image (i.e. a shipping container ID in a logistics use case) for further processing. You can then use a DynamicCropBlock to crop the region of interest before running OCR.
Using a detections model then cropping detections allows you to isolate your analysis on particular regions of an image.
Type identifier¶
Use the following identifier in step "type"
field: roboflow_core/ocr_model@v1
to add the block as
as step in your workflow.
Properties¶
Name | Type | Description | Refs |
---|---|---|---|
name |
str |
Unique name of step in workflows. | ❌ |
The Refs column marks possibility to parametrise the property with dynamic values available
in workflow
runtime. See Bindings for more info.
Available Connections¶
Compatible Blocks
Check what blocks you can connect to OCR Model
in version v1
.
- inputs:
Depth Estimation
,Classification Label Visualization
,Stability AI Image Generation
,Camera Calibration
,Ellipse Visualization
,SIFT
,Blur Visualization
,Circle Visualization
,Dot Visualization
,Image Blur
,Background Color Visualization
,Stitch Images
,Image Slicer
,Absolute Static Crop
,Color Visualization
,Pixelate Visualization
,Label Visualization
,Stability AI Inpainting
,Polygon Zone Visualization
,Relative Static Crop
,Keypoint Visualization
,Bounding Box Visualization
,Mask Visualization
,Model Comparison Visualization
,Image Preprocessing
,Line Counter Visualization
,Grid Visualization
,SIFT Comparison
,Image Convert Grayscale
,Halo Visualization
,Trace Visualization
,Corner Visualization
,Polygon Visualization
,Triangle Visualization
,Image Contours
,Perspective Correction
,Dynamic Crop
,Crop Visualization
,Camera Focus
,Reference Path Visualization
,Image Slicer
,Image Threshold
- outputs:
LMM
,Classification Label Visualization
,LMM For Classification
,Clip Comparison
,Line Counter
,Instance Segmentation Model
,CogVLM
,Roboflow Dataset Upload
,Distance Measurement
,Twilio SMS Notification
,Florence-2 Model
,OpenAI
,Label Visualization
,Path Deviation
,Roboflow Dataset Upload
,Detections Stitch
,Bounding Box Visualization
,Llama 3.2 Vision
,Model Comparison Visualization
,Slack Notification
,Pixel Color Count
,Halo Visualization
,Triangle Visualization
,Model Monitoring Inference Aggregator
,Time in Zone
,Reference Path Visualization
,Perspective Correction
,Dynamic Crop
,Time in Zone
,Cache Get
,Florence-2 Model
,Cache Set
,Local File Sink
,Path Deviation
,OpenAI
,Google Vision OCR
,Stability AI Image Generation
,Ellipse Visualization
,Size Measurement
,Circle Visualization
,Dot Visualization
,Image Blur
,CLIP Embedding Model
,Background Color Visualization
,Color Visualization
,Google Gemini
,Stability AI Inpainting
,Polygon Zone Visualization
,Keypoint Visualization
,Mask Visualization
,Image Preprocessing
,Line Counter Visualization
,Roboflow Custom Metadata
,Anthropic Claude
,Webhook Sink
,SIFT Comparison
,YOLO-World Model
,Trace Visualization
,Instance Segmentation Model
,Corner Visualization
,Polygon Visualization
,Segment Anything 2 Model
,Crop Visualization
,Email Notification
,Line Counter
,Image Threshold
Input and Output Bindings¶
The available connections depend on its binding kinds. Check what binding kinds
OCR Model
in version v1
has.
Bindings
-
input
images
(image
): The image to infer on..
-
output
result
(string
): String value.parent_id
(parent_id
): Identifier of parent for step output.root_parent_id
(parent_id
): Identifier of parent for step output.prediction_type
(prediction_type
): String value with type of prediction.
Example JSON definition of step OCR Model
in version v1
{
"name": "<your_step_name_here>",
"type": "roboflow_core/ocr_model@v1",
"images": "$inputs.image"
}