Template Matching¶
Class: TemplateMatchingBlockV1
Source: inference.core.workflows.core_steps.classical_cv.template_matching.v1.TemplateMatchingBlockV1
Locate instances of a template image within a larger image using template matching with normalized cross-correlation, finding exact or near-exact matches of the template pattern at any location in the image, outputting bounding box detections with optional NMS filtering for object detection, logo detection, pattern recognition, and template-based object localization workflows.
How This Block Works¶
This block searches for occurrences of a template image within a larger input image using normalized cross-correlation template matching. The block:
- Receives an input image and a template image (smaller pattern to search for)
- Converts both images to grayscale for template matching (template matching typically works on grayscale images for efficiency and robustness)
- Performs template matching using OpenCV's matchTemplate with TM_CCOEFF_NORMED method:
- Slides the template across the input image at every possible position
- Computes normalized cross-correlation coefficient at each position (measures similarity between template and image region)
- Generates a similarity map showing how well the template matches at each location
- Identifies match locations where similarity exceeds the matching_threshold:
- Finds all positions where the correlation coefficient is greater than or equal to the threshold
- Threshold values range from 0.0 to 1.0, with higher values requiring closer matches
- Lower thresholds find more potential matches (including partial matches), higher thresholds find only very similar matches
- Creates bounding boxes for each match:
- Each match location becomes a detection with a bounding box matching the template's dimensions
- All detections have confidence of 1.0 (they met the threshold requirement)
- All detections are assigned class "template_match" and class_id 0
- Each detection gets a unique detection ID for tracking
- Optionally applies Non-Maximum Suppression (NMS) to filter overlapping detections:
- Template matching often produces many overlapping detections at the same location (duplicate matches)
- NMS removes overlapping detections, keeping only the best match in each area
- NMS threshold controls how much overlap is allowed before removing detections
- Can be disabled (apply_nms=False) if NMS becomes computationally intractable with very large numbers of matches
- Attaches metadata to detections:
- Sets parent_id to reference the input image
- Sets prediction_type to "object-detection"
- Stores image dimensions for coordinate reference
- Attaches parent coordinate information for workflow tracking
- Returns detection predictions in sv.Detections format along with the total number of matches found
The block uses normalized cross-correlation which is effective for finding exact or near-exact template matches. It works best when the template appears in the image at the same scale, rotation, and lighting conditions. The method tends to produce many overlapping detections for the same match location, which is why NMS filtering is important. However, in cases with extremely large numbers of matches (e.g., repeating patterns), NMS may become computationally expensive and can be disabled if needed.
Common Use Cases¶
- Logo and Brand Detection: Find specific logos or brand elements within images (e.g., detect company logos in photos, find brand markers in images, locate specific logo patterns in scenes), enabling logo detection workflows
- Exact Pattern Matching: Locate specific patterns or objects that appear identically in images (e.g., find specific UI elements in screenshots, detect exact patterns in images, locate specific visual elements), enabling exact pattern detection workflows
- Quality Control and Inspection: Find reference patterns or features for quality inspection (e.g., detect specific features in manufacturing images, find reference markers for alignment, locate inspection targets), enabling quality control workflows
- Object Localization: Locate specific objects or regions when exact appearance is known (e.g., find specific objects with known appearance, locate reference objects in images, detect specific visual elements), enabling template-based object localization
- Document Processing: Find specific elements or regions in documents (e.g., locate form fields in documents, detect specific document elements, find reference markers in scanned documents), enabling document processing workflows
- UI Element Detection: Detect specific UI components or elements in interface images (e.g., find buttons in UI screenshots, locate specific UI elements, detect interface components), enabling UI analysis workflows
Connecting to Other Blocks¶
This block receives an image and template, and produces detection predictions:
- After image input blocks to find template patterns in input images (e.g., search for templates in input images, locate patterns in camera feeds, find templates in image streams), enabling template matching workflows
- After preprocessing blocks to find templates in preprocessed images (e.g., match templates after image enhancement, find patterns in filtered images, locate templates in normalized images), enabling preprocessed template matching
- Before visualization blocks to visualize template match locations (e.g., visualize detected template matches, display bounding boxes for matches, show template match results), enabling template match visualization workflows
- Before filtering blocks to filter template matches by criteria (e.g., filter matches by location, select specific match regions, refine template match results), enabling filtered template matching workflows
- Before crop blocks to extract regions around template matches (e.g., crop areas around matches, extract match regions for analysis, crop template match locations), enabling template-based region extraction
- In quality control workflows where template matching is used for inspection or alignment (e.g., find reference markers for alignment, detect inspection targets, locate quality control features), enabling quality control template matching workflows
Type identifier¶
Use the following identifier in step "type" field: roboflow_core/template_matching@v1to add the block as
as step in your workflow.
Properties¶
| Name | Type | Description | Refs |
|---|---|---|---|
name |
str |
Enter a unique identifier for this step.. | ❌ |
matching_threshold |
float |
Minimum similarity threshold (0.0 to 1.0) required for a template match. Higher values (closer to 1.0) require very close matches and find fewer but more precise matches. Lower values (closer to 0.0) allow more lenient matches and find more potential matches including partial matches. Default is 0.8, which requires fairly close matches. Use lower thresholds (0.6-0.7) to find more matches or handle slight variations. Use higher thresholds (0.85-0.95) for exact matches only. The threshold compares normalized cross-correlation coefficients from template matching.. | ✅ |
apply_nms |
bool |
Whether to apply Non-Maximum Suppression (NMS) to filter overlapping detections. Template matching often produces many overlapping detections at the same location. NMS removes overlapping detections, keeping only the best match in each area. Default is True (recommended for most cases). Set to False if: (1) the number of matches is extremely large (NMS may become computationally expensive), (2) you want to see all raw matches without filtering, or (3) matches are intentionally close together and should all be kept. When disabled, you may see many duplicate detections for the same match location.. | ✅ |
nms_threshold |
float |
Intersection over Union (IoU) threshold for Non-Maximum Suppression. Only relevant when apply_nms is True. Detections with IoU overlap greater than this threshold are considered duplicates, and only the detection with highest confidence is kept. Lower values (0.3-0.4) are more aggressive at removing overlaps, removing detections that are only slightly overlapping. Higher values (0.6-0.7) are more lenient, only removing heavily overlapping detections. Default is 0.5, which provides balanced overlap filtering. Adjust based on how much overlap you expect between template matches and how close together valid matches can be.. | ✅ |
The Refs column marks possibility to parametrise the property with dynamic values available
in workflow runtime. See Bindings for more info.
Available Connections¶
Compatible Blocks
Check what blocks you can connect to Template Matching in version v1.
- inputs:
Roboflow Dataset Upload,Line Counter Visualization,Stability AI Outpainting,Email Notification,Gaze Detection,Image Slicer,Identify Outliers,Image Preprocessing,Color Visualization,Cosine Similarity,Ellipse Visualization,Polygon Visualization,Relative Static Crop,Detections Consensus,Webhook Sink,Model Comparison Visualization,Trace Visualization,Camera Focus,Roboflow Custom Metadata,VLM As Classifier,Image Threshold,Stitch Images,Heatmap Visualization,SIFT Comparison,Morphological Transformation,Halo Visualization,Crop Visualization,Camera Calibration,Dot Visualization,S3 Sink,Twilio SMS Notification,Icon Visualization,Model Monitoring Inference Aggregator,Local File Sink,Roboflow Dataset Upload,Dynamic Zone,Image Contours,VLM As Classifier,JSON Parser,Pixelate Visualization,Twilio SMS/MMS Notification,Polygon Zone Visualization,Reference Path Visualization,Motion Detection,Blur Visualization,Background Subtraction,Text Display,Clip Comparison,VLM As Detector,Stability AI Image Generation,Perspective Correction,Bounding Box Visualization,Depth Estimation,Identify Changes,Classification Label Visualization,Image Slicer,Absolute Static Crop,Image Blur,Stability AI Inpainting,Polygon Visualization,Image Convert Grayscale,SIFT,Roboflow Vision Events,VLM As Detector,Label Visualization,Corner Visualization,Grid Visualization,Dynamic Crop,Contrast Equalization,Keypoint Visualization,Triangle Visualization,QR Code Generator,Halo Visualization,Circle Visualization,Camera Focus,Mask Visualization,Morphological Transformation,Contrast Enhancement,Background Color Visualization,Email Notification,PTZ Tracking (ONVIF),Slack Notification,SIFT Comparison - outputs:
Roboflow Dataset Upload,Line Counter Visualization,Mask Edge Snap,Image Slicer,Distance Measurement,Instance Segmentation Model,Color Visualization,Ellipse Visualization,ByteTrack Tracker,Polygon Visualization,Byte Tracker,Detections Consensus,Detections Classes Replacement,Webhook Sink,Stitch OCR Detections,Trace Visualization,Camera Focus,Object Detection Model,Size Measurement,Image Threshold,Heatmap Visualization,SORT Tracker,Florence-2 Model,Halo Visualization,Detections Transformation,Path Deviation,Dot Visualization,Path Deviation,Twilio SMS Notification,Model Monitoring Inference Aggregator,Roboflow Dataset Upload,Dynamic Zone,Pixelate Visualization,Line Counter,Twilio SMS/MMS Notification,Motion Detection,Blur Visualization,Background Subtraction,Text Display,Detections Merge,Perspective Correction,Overlap Filter,Line Counter,Velocity,Bounding Box Visualization,Anthropic Claude,Stability AI Inpainting,Polygon Visualization,Roboflow Vision Events,Label Visualization,Grid Visualization,Per-Class Confidence Filter,Triangle Visualization,Halo Visualization,Circle Visualization,Segment Anything 2 Model,Mask Visualization,Dominant Color,Email Notification,Slack Notification,Detections Stabilizer,Detections Stitch,Object Detection Model,Stability AI Outpainting,Email Notification,Identify Outliers,Image Preprocessing,Detections Combine,Object Detection Model,SAM2 Video Tracker,Detection Event Log,Byte Tracker,Anthropic Claude,Time in Zone,Model Comparison Visualization,Roboflow Custom Metadata,Detection Offset,Instance Segmentation Model,Detections List Roll-Up,Mask Area Measurement,Stitch Images,SIFT Comparison,Morphological Transformation,Instance Segmentation Model,Crop Visualization,Florence-2 Model,Time in Zone,OC-SORT Tracker,Icon Visualization,Detections Filter,Image Contours,Keypoint Detection Model,Time in Zone,Reference Path Visualization,Anthropic Claude,Pixel Color Count,Identify Changes,Classification Label Visualization,Byte Tracker,Absolute Static Crop,Image Blur,Image Slicer,Corner Visualization,Dynamic Crop,Keypoint Detection Model,Keypoint Visualization,QR Code Generator,Morphological Transformation,Keypoint Detection Model,Background Color Visualization,PTZ Tracking (ONVIF),Stitch OCR Detections,SIFT Comparison
Input and Output Bindings¶
The available connections depend on its binding kinds. Check what binding kinds
Template Matching in version v1 has.
Bindings
-
input
image(image): Large image in which to search for the template pattern. The template will be searched across this entire image at all possible positions. The image is converted to grayscale internally for template matching. Template matching works best when the image and template have similar lighting conditions and the template appears at similar scale and orientation in the image..template(image): Small template image pattern to search for within the input image. The template should be smaller than the input image. The template is converted to grayscale internally for matching. Template matching finds exact or near-exact matches of this template at any location in the input image. Works best when the template appears in the image at the same scale, rotation, and lighting conditions. The template's dimensions determine the size of the detection bounding boxes..matching_threshold(float): Minimum similarity threshold (0.0 to 1.0) required for a template match. Higher values (closer to 1.0) require very close matches and find fewer but more precise matches. Lower values (closer to 0.0) allow more lenient matches and find more potential matches including partial matches. Default is 0.8, which requires fairly close matches. Use lower thresholds (0.6-0.7) to find more matches or handle slight variations. Use higher thresholds (0.85-0.95) for exact matches only. The threshold compares normalized cross-correlation coefficients from template matching..apply_nms(boolean): Whether to apply Non-Maximum Suppression (NMS) to filter overlapping detections. Template matching often produces many overlapping detections at the same location. NMS removes overlapping detections, keeping only the best match in each area. Default is True (recommended for most cases). Set to False if: (1) the number of matches is extremely large (NMS may become computationally expensive), (2) you want to see all raw matches without filtering, or (3) matches are intentionally close together and should all be kept. When disabled, you may see many duplicate detections for the same match location..nms_threshold(float_zero_to_one): Intersection over Union (IoU) threshold for Non-Maximum Suppression. Only relevant when apply_nms is True. Detections with IoU overlap greater than this threshold are considered duplicates, and only the detection with highest confidence is kept. Lower values (0.3-0.4) are more aggressive at removing overlaps, removing detections that are only slightly overlapping. Higher values (0.6-0.7) are more lenient, only removing heavily overlapping detections. Default is 0.5, which provides balanced overlap filtering. Adjust based on how much overlap you expect between template matches and how close together valid matches can be..
-
output
predictions(object_detection_prediction): Prediction with detected bounding boxes in form of sv.Detections(...) object.number_of_matches(integer): Integer value.
Example JSON definition of step Template Matching in version v1
{
"name": "<your_step_name_here>",
"type": "roboflow_core/template_matching@v1",
"image": "$inputs.image",
"template": "$inputs.template",
"matching_threshold": 0.8,
"apply_nms": "$inputs.apply_nms",
"nms_threshold": "$inputs.nms_threshold"
}