Google Vision OCR¶
Class: GoogleVisionOCRBlockV1
Source: inference.core.workflows.core_steps.models.foundation.google_vision_ocr.v1.GoogleVisionOCRBlockV1
Detect text in images using Google Vision OCR.
Supported types of text detection:
text_detection: optimized for areas of text within a larger image.ocr_text_detection: optimized for dense text documents.
Provide your Google Vision API key or set the value to rf_key:account (or
rf_key:user:<id>) to proxy requests through Roboflow's API.
Type identifier¶
Use the following identifier in step "type" field: roboflow_core/google_vision_ocr@v1to add the block as
as step in your workflow.
Properties¶
| Name | Type | Description | Refs |
|---|---|---|---|
name |
str |
Enter a unique identifier for this step.. | ❌ |
ocr_type |
str |
Type of OCR to use. | ❌ |
api_key |
str |
Your Google Vision API key. | ✅ |
language_hints |
List[str] |
Optional list of language codes to pass to the OCR API. If not provided, the API will attempt to detect the language automatically.If provided, language codes must be supported by the OCR API, visit https://cloud.google.com/vision/docs/languages for list of supported language codes.. | ❌ |
The Refs column marks possibility to parametrise the property with dynamic values available
in workflow runtime. See Bindings for more info.
Available Connections¶
Compatible Blocks
Check what blocks you can connect to Google Vision OCR in version v1.
- inputs:
Dynamic Crop,OCR Model,Email Notification,Image Blur,Background Subtraction,SIFT Comparison,OpenAI,Google Vision OCR,Google Gemini,Image Preprocessing,Instance Segmentation Model,Local File Sink,Single-Label Classification Model,Bounding Box Visualization,Model Monitoring Inference Aggregator,Anthropic Claude,Multi-Label Classification Model,Keypoint Detection Model,Email Notification,Slack Notification,Camera Focus,Twilio SMS/MMS Notification,Dot Visualization,Florence-2 Model,Roboflow Dataset Upload,CSV Formatter,Camera Focus,Stitch OCR Detections,Depth Estimation,Polygon Visualization,Perspective Correction,OpenAI,Camera Calibration,Corner Visualization,Icon Visualization,Image Slicer,Qwen3.5-VL,Line Counter Visualization,Heatmap Visualization,Morphological Transformation,Stability AI Image Generation,Google Gemini,Keypoint Visualization,VLM As Detector,Halo Visualization,Background Color Visualization,Label Visualization,Polygon Visualization,Pixelate Visualization,LMM,CogVLM,Contrast Equalization,Triangle Visualization,Stability AI Outpainting,Mask Visualization,VLM As Classifier,Color Visualization,Text Display,Relative Static Crop,Reference Path Visualization,Stitch OCR Detections,Llama 3.2 Vision,OpenAI,Image Threshold,Clip Comparison,Classification Label Visualization,Webhook Sink,Circle Visualization,Polygon Zone Visualization,Image Contours,Image Convert Grayscale,Grid Visualization,Florence-2 Model,Roboflow Custom Metadata,LMM For Classification,SIFT,Halo Visualization,Object Detection Model,Anthropic Claude,Google Gemini,Model Comparison Visualization,Blur Visualization,QR Code Generator,EasyOCR,Absolute Static Crop,Image Slicer,S3 Sink,Anthropic Claude,Stability AI Inpainting,Ellipse Visualization,Crop Visualization,Trace Visualization,Twilio SMS Notification,Stitch Images,OpenAI,Roboflow Dataset Upload - outputs:
Dynamic Crop,Image Blur,Google Vision OCR,Google Gemini,Image Preprocessing,Local File Sink,Bounding Box Visualization,Model Monitoring Inference Aggregator,Dot Visualization,Florence-2 Model,Roboflow Dataset Upload,Depth Estimation,Polygon Visualization,OpenAI,Line Counter,Detections List Roll-Up,Line Counter Visualization,Heatmap Visualization,Google Gemini,Stability AI Image Generation,Morphological Transformation,Distance Measurement,Keypoint Visualization,Background Color Visualization,Label Visualization,Polygon Visualization,LMM,CogVLM,Time in Zone,Triangle Visualization,Stability AI Outpainting,Mask Visualization,Color Visualization,Detections Combine,Text Display,Reference Path Visualization,OpenAI,Llama 3.2 Vision,Image Threshold,Clip Comparison,Classification Label Visualization,Polygon Zone Visualization,Roboflow Custom Metadata,LMM For Classification,Velocity,Halo Visualization,Blur Visualization,Path Deviation,Anthropic Claude,SAM 3,Detections Transformation,Ellipse Visualization,Crop Visualization,Path Deviation,Trace Visualization,Twilio SMS Notification,Size Measurement,Detections Stabilizer,Detections Merge,Time in Zone,Email Notification,OpenAI,SIFT Comparison,Seg Preview,Time in Zone,Instance Segmentation Model,Anthropic Claude,Email Notification,Slack Notification,Twilio SMS/MMS Notification,Detections Stitch,Cache Set,Camera Focus,SAM 3,Stitch OCR Detections,Perspective Correction,PTZ Tracking (ONVIF),Moondream2,Icon Visualization,Corner Visualization,Overlap Filter,Byte Tracker,Halo Visualization,Detection Event Log,Pixelate Visualization,Contrast Equalization,Instance Segmentation Model,Detections Classes Replacement,Line Counter,Stitch OCR Detections,Webhook Sink,Circle Visualization,Mask Area Measurement,Byte Tracker,Florence-2 Model,SAM 3,Perception Encoder Embedding Model,Cache Get,YOLO-World Model,Byte Tracker,Detections Consensus,Anthropic Claude,Google Gemini,Model Comparison Visualization,Detection Offset,QR Code Generator,S3 Sink,CLIP Embedding Model,Stability AI Inpainting,Segment Anything 2 Model,OpenAI,Pixel Color Count,Detections Filter,Roboflow Dataset Upload
Input and Output Bindings¶
The available connections depend on its binding kinds. Check what binding kinds
Google Vision OCR in version v1 has.
Bindings
-
input
image(image): Image to run OCR.api_key(Union[string,ROBOFLOW_MANAGED_KEY,secret]): Your Google Vision API key.
-
output
text(string): String value.language(string): String value.predictions(object_detection_prediction): Prediction with detected bounding boxes in form of sv.Detections(...) object.
Example JSON definition of step Google Vision OCR in version v1
{
"name": "<your_step_name_here>",
"type": "roboflow_core/google_vision_ocr@v1",
"image": "$inputs.image",
"ocr_type": "<block_does_not_provide_example>",
"api_key": "xxx-xxx",
"language_hints": [
"en",
"fr"
]
}