Line Counter¶
v2¶
Class: LineCounterBlockV2 (there are multiple versions of this block)
Source: inference.core.workflows.core_steps.analytics.line_counter.v2.LineCounterBlockV2
Warning: This block has multiple versions. Please refer to the specific version for details. You can learn more about how versions work here: Versioning
Count objects crossing a defined line segment in video using tracked detections, maintaining separate counts for objects crossing in opposite directions (in and out), and outputting both count values and the actual detection objects that crossed the line for traffic analysis, people counting, entry/exit monitoring, and directional flow measurement workflows.
How This Block Works¶
This block counts objects that cross a line segment by tracking their movement across video frames. The block:
- Receives tracked detection predictions with unique tracker IDs and an image with embedded video metadata
- Extracts video metadata from the image:
- Accesses video_metadata from the WorkflowImageData object
- Extracts video_identifier to maintain separate counting state for different videos
- Uses video metadata to initialize and manage line zone state per video
- Validates that detections have tracker IDs (required for tracking object movement across frames)
- Initializes or retrieves a line zone for the video:
- Creates a LineZone from two coordinate points defining the line segment
- Configures triggering anchor point if specified (optional - if not specified, uses default anchor behavior)
- Stores line zone configuration per video using video_identifier
- Maintains separate counting state for each video
- Monitors object positions across frames:
- Tracks each object's position using its unique tracker_id
- Detects when an object's triggering anchor point (if specified) or default anchor crosses the line
- Determines crossing direction based on which side of the line the object approaches from
- Counts line crossings:
- In Direction: Objects crossing the line in one direction increment the count_in counter
- Out Direction: Objects crossing the line in the opposite direction increment the count_out counter
- Each unique tracker_id is counted only once per crossing (prevents duplicate counting if object oscillates near line)
- Identifies crossing detections:
- Creates masks identifying which detections crossed in each direction in the current frame
- Filters detections to separate those that crossed "in" from those that crossed "out"
- Returns the actual detection objects (not just counts) for further processing
- Maintains persistent counting state:
- Counts accumulate across frames for the entire video
- State persists for each video until workflow execution completes
- Separate counters for each unique video_identifier
- Returns four outputs:
- count_in: Total number of objects that crossed the line in the "in" direction (cumulative across video)
- count_out: Total number of objects that crossed the line in the "out" direction (cumulative across video)
- detections_in: Detection objects that crossed the line in the "in" direction (current frame crossings)
- detections_out: Detection objects that crossed the line in the "out" direction (current frame crossings)
The line segment defines a virtual boundary in the video frame. The direction (in/out) is determined by which side of the line objects approach from - for a horizontal line, objects coming from above might count as "in" while objects from below count as "out" (or vice versa, depending on line orientation). The triggering anchor (if specified) determines which point on the bounding box must cross the line for the crossing to be counted - if not specified, the line zone uses its default anchor behavior. The count outputs provide cumulative totals across the video, while the detection outputs provide the actual objects that crossed in the current frame, enabling further analysis or visualization of crossing events.
Common Use Cases¶
- People Counting: Count people entering and exiting buildings, stores, or events (e.g., count visitors entering store, track people entering/exiting building, monitor event attendance), enabling entry/exit counting workflows
- Traffic Analysis: Count vehicles passing through intersections or road segments (e.g., count vehicles crossing intersection, track traffic flow in specific directions, monitor vehicle passage at checkpoints), enabling traffic flow analysis workflows
- Retail Analytics: Track customer movement and foot traffic in retail spaces (e.g., count customers entering store sections, track movement between departments, monitor shopping flow patterns), enabling retail foot traffic analytics workflows
- Security Monitoring: Monitor entry and exit at secure areas or checkpoints (e.g., track entries to restricted areas, count people at access points, monitor checkpoint crossings), enabling security access monitoring workflows
- Occupancy Management: Track occupancy changes by counting objects entering and leaving spaces (e.g., count entries/exits to manage room capacity, track vehicle arrivals/departures in parking, monitor space occupancy changes), enabling occupancy tracking workflows
- Wildlife Monitoring: Count animals crossing defined paths or boundaries (e.g., track animal migration patterns, count wildlife crossing roads, monitor animal movement in habitats), enabling wildlife behavior analysis workflows
Connecting to Other Blocks¶
This block receives tracked detections and an image with embedded video metadata, and produces count_in, count_out, detections_in, and detections_out:
- After Byte Tracker blocks to count tracked objects crossing lines (e.g., count tracked people crossing line, track vehicle crossings with consistent IDs, monitor tracked object movements), enabling tracking-to-counting workflows
- After object detection or instance segmentation blocks with tracking enabled to count detected objects (e.g., count detected vehicles, track people crossings, monitor object movements), enabling detection-to-counting workflows
- Using detections_in or detections_out outputs to process or visualize objects that crossed the line (e.g., visualize objects that crossed, analyze crossing objects, filter for crossing events), enabling crossing object analysis workflows
- Before visualization blocks to display line counter information and crossing objects (e.g., visualize line and counts, display crossing statistics, show crossing objects with annotations), enabling counting visualization workflows
- Before data storage blocks to record counting data and crossing events (e.g., log entry/exit counts, store traffic statistics, record crossing objects with metadata), enabling counting data logging workflows
- Before notification blocks to alert on count thresholds or crossing events (e.g., alert when count exceeds limit, notify on specific object crossings, trigger actions based on counts), enabling count-based notification workflows
Version Differences¶
Enhanced from v1:
- Detection Outputs: Adds two new outputs (
detections_inanddetections_out) that provide the actual detection objects that crossed the line in each direction, not just count totals, enabling downstream processing and visualization of crossing objects - Simplified Input: Uses
imageinput that contains embedded video metadata instead of requiring a separatemetadatafield, simplifying workflow connections and reducing input complexity - Optional Triggering Anchor: Makes
triggering_anchoroptional (default None) instead of required, allowing the line zone to use its default anchor behavior when no specific anchor is needed - Improved Integration: Better integration with image-based workflows since video metadata is accessed directly from the image object rather than requiring separate metadata input
Requirements¶
This block requires tracked detections with tracker_id information (detections must come from a tracking block like Byte Tracker). The line must be defined as a list of exactly 2 points, where each point is a list or tuple of exactly 2 coordinates (x, y). The image's video_metadata should include video_identifier to maintain separate counting state for different videos. The block maintains persistent counting state across frames for each video, so it should be used in video workflows where frames are processed sequentially. For accurate counting, detections should be provided consistently across frames with valid tracker IDs.
Type identifier¶
Use the following identifier in step "type" field: roboflow_core/line_counter@v2to add the block as
as step in your workflow.
Properties¶
| Name | Type | Description | Refs |
|---|---|---|---|
name |
str |
Enter a unique identifier for this step.. | ❌ |
line_segment |
List[Any] |
Line segment defined by exactly two points, each with [x, y] coordinates. Objects crossing from one side count as 'in', objects crossing from the other side count as 'out'. Example: [[0, 100], [500, 100]] creates a horizontal line at y=100. Crossing direction depends on which side objects approach from.. | ✅ |
triggering_anchor |
str |
Optional point on the bounding box that must cross the line for counting. If not specified (None), the line zone uses its default anchor behavior. Options when specified: CENTER, BOTTOM_CENTER, TOP_CENTER, CENTER_LEFT, CENTER_RIGHT, etc. Specifying CENTER ensures the object is substantially across the line before counting, reducing false positives from objects near but not fully crossing the line.. | ✅ |
The Refs column marks possibility to parametrise the property with dynamic values available
in workflow runtime. See Bindings for more info.
Available Connections¶
Compatible Blocks
Check what blocks you can connect to Line Counter in version v2.
- inputs:
Detections Stitch,Detections Stabilizer,Roboflow Dataset Upload,Object Detection Model,Email Notification,Google Gemma API,Mask Edge Snap,OCR Model,Google Vision OCR,Instance Segmentation Model,Google Gemini,EasyOCR,Detections Combine,Object Detection Model,SAM2 Video Tracker,Detection Event Log,Byte Tracker,ByteTrack Tracker,Bounding Rectangle,OpenAI,Anthropic Claude,Byte Tracker,Detections Consensus,Detections Classes Replacement,Time in Zone,Webhook Sink,Stitch OCR Detections,Object Detection Model,Qwen 3.5 API,Camera Focus,Roboflow Custom Metadata,YOLO-World Model,OpenAI,Detection Offset,Buffer,SAM 3,Instance Segmentation Model,Single-Label Classification Model,Detections List Roll-Up,Size Measurement,VLM As Classifier,Template Matching,Mask Area Measurement,Qwen 3.6 API,SORT Tracker,Florence-2 Model,Detections Transformation,Instance Segmentation Model,CogVLM,Florence-2 Model,Path Deviation,Time in Zone,GLM-OCR,OC-SORT Tracker,S3 Sink,Path Deviation,SAM 3,Seg Preview,Twilio SMS Notification,Model Monitoring Inference Aggregator,Detections Filter,Google Gemini,Local File Sink,Roboflow Dataset Upload,Dynamic Zone,Clip Comparison,Line Counter,Twilio SMS/MMS Notification,Time in Zone,Motion Detection,Dimension Collapse,Anthropic Claude,Clip Comparison,CSV Formatter,VLM As Detector,LMM,Detections Merge,Perspective Correction,Overlap Filter,Anthropic Claude,Velocity,Multi-Label Classification Model,Byte Tracker,SAM 3,Roboflow Vision Events,VLM As Detector,OpenAI,Google Gemini,Dynamic Crop,Per-Class Confidence Filter,Moondream2,Qwen3.5-VL,Segment Anything 2 Model,LMM For Classification,OpenAI,Keypoint Detection Model,MoonshotAI Kimi,Llama 3.2 Vision,Email Notification,PTZ Tracking (ONVIF),Slack Notification,Stitch OCR Detections - outputs:
Roboflow Dataset Upload,Line Counter Visualization,Mask Edge Snap,Image Slicer,Instance Segmentation Model,Distance Measurement,Color Visualization,Bounding Rectangle,Ellipse Visualization,ByteTrack Tracker,Polygon Visualization,Byte Tracker,Detections Consensus,Detections Classes Replacement,Webhook Sink,Trace Visualization,Object Detection Model,Stitch OCR Detections,Camera Focus,Size Measurement,Image Threshold,Heatmap Visualization,SORT Tracker,Florence-2 Model,Halo Visualization,Detections Transformation,Path Deviation,Dot Visualization,Path Deviation,Twilio SMS Notification,Model Monitoring Inference Aggregator,Roboflow Dataset Upload,Dynamic Zone,Pixelate Visualization,Line Counter,Twilio SMS/MMS Notification,Motion Detection,Blur Visualization,Background Subtraction,Text Display,Detections Merge,Perspective Correction,Anthropic Claude,Overlap Filter,Bounding Box Visualization,Line Counter,Velocity,Stability AI Inpainting,Polygon Visualization,Roboflow Vision Events,Label Visualization,Grid Visualization,Per-Class Confidence Filter,Triangle Visualization,Halo Visualization,Circle Visualization,Segment Anything 2 Model,Mask Visualization,Dominant Color,Email Notification,Slack Notification,Detections Stabilizer,Detections Stitch,Object Detection Model,Stability AI Outpainting,Email Notification,Identify Outliers,Image Preprocessing,Detections Combine,Object Detection Model,SAM2 Video Tracker,Detection Event Log,Byte Tracker,Anthropic Claude,Time in Zone,Model Comparison Visualization,Roboflow Custom Metadata,Detection Offset,Instance Segmentation Model,Detections List Roll-Up,Mask Area Measurement,Stitch Images,SIFT Comparison,Morphological Transformation,Instance Segmentation Model,Crop Visualization,Florence-2 Model,Time in Zone,OC-SORT Tracker,Icon Visualization,Detections Filter,Image Contours,Keypoint Detection Model,Time in Zone,Reference Path Visualization,Anthropic Claude,Pixel Color Count,Identify Changes,Classification Label Visualization,Byte Tracker,Absolute Static Crop,Image Blur,Image Slicer,Corner Visualization,Keypoint Detection Model,Dynamic Crop,Keypoint Visualization,QR Code Generator,Morphological Transformation,Keypoint Detection Model,Background Color Visualization,PTZ Tracking (ONVIF),Stitch OCR Detections,SIFT Comparison
Input and Output Bindings¶
The available connections depend on its binding kinds. Check what binding kinds
Line Counter in version v2 has.
Bindings
-
input
image(image): Image with embedded video metadata. The video_metadata contains video_identifier to maintain separate counting state for different videos. Required for persistent counting across frames..detections(Union[object_detection_prediction,instance_segmentation_prediction]): Tracked object detection or instance segmentation predictions. Must include tracker_id information from a tracking block. Objects are counted when their triggering anchor point (if specified) crosses the line segment. The detections_in and detections_out outputs provide the actual detection objects that crossed in each direction..line_segment(list_of_values): Line segment defined by exactly two points, each with [x, y] coordinates. Objects crossing from one side count as 'in', objects crossing from the other side count as 'out'. Example: [[0, 100], [500, 100]] creates a horizontal line at y=100. Crossing direction depends on which side objects approach from..triggering_anchor(string): Optional point on the bounding box that must cross the line for counting. If not specified (None), the line zone uses its default anchor behavior. Options when specified: CENTER, BOTTOM_CENTER, TOP_CENTER, CENTER_LEFT, CENTER_RIGHT, etc. Specifying CENTER ensures the object is substantially across the line before counting, reducing false positives from objects near but not fully crossing the line..
-
output
count_in(integer): Integer value.count_out(integer): Integer value.detections_in(Union[object_detection_prediction,instance_segmentation_prediction]): Prediction with detected bounding boxes in form of sv.Detections(...) object ifobject_detection_predictionor Prediction with detected bounding boxes and segmentation masks in form of sv.Detections(...) object ifinstance_segmentation_prediction.detections_out(Union[object_detection_prediction,instance_segmentation_prediction]): Prediction with detected bounding boxes in form of sv.Detections(...) object ifobject_detection_predictionor Prediction with detected bounding boxes and segmentation masks in form of sv.Detections(...) object ifinstance_segmentation_prediction.
Example JSON definition of step Line Counter in version v2
{
"name": "<your_step_name_here>",
"type": "roboflow_core/line_counter@v2",
"image": "<block_does_not_provide_example>",
"detections": "$steps.object_detection_model.predictions",
"line_segment": [
[
0,
50
],
[
500,
50
]
],
"triggering_anchor": "CENTER"
}
v1¶
Class: LineCounterBlockV1 (there are multiple versions of this block)
Source: inference.core.workflows.core_steps.analytics.line_counter.v1.LineCounterBlockV1
Warning: This block has multiple versions. Please refer to the specific version for details. You can learn more about how versions work here: Versioning
Count objects crossing a defined line segment in video using tracked detections, maintaining separate counts for objects crossing in opposite directions (in and out) for traffic analysis, people counting, entry/exit monitoring, and directional flow measurement workflows.
How This Block Works¶
This block counts objects that cross a line segment by tracking their movement across video frames. The block:
- Receives tracked detection predictions with unique tracker IDs and video metadata
- Validates that detections have tracker IDs (required for tracking object movement across frames)
- Initializes or retrieves a line zone for the video:
- Creates a LineZone from two coordinate points defining the line segment
- Stores line zone configuration per video using video_identifier
- Maintains separate counting state for each video
- Monitors object positions across frames:
- Tracks each object's position using its unique tracker_id
- Detects when an object's triggering anchor point (default: CENTER of bounding box) crosses the line
- Determines crossing direction based on which side of the line the object approaches from
- Counts line crossings:
- In Direction: Objects crossing the line in one direction increment the count_in counter
- Out Direction: Objects crossing the line in the opposite direction increment the count_out counter
- Each unique tracker_id is counted only once per crossing (prevents duplicate counting if object oscillates near line)
- Maintains persistent counting state:
- Counts accumulate across frames for the entire video
- State persists for each video until workflow execution completes
- Separate counters for each unique video_identifier
- Returns two count values:
- count_in: Total number of objects that crossed the line in the "in" direction
- count_out: Total number of objects that crossed the line in the "out" direction
The line segment defines a virtual boundary in the video frame. The direction (in/out) is determined by which side of the line objects approach from - for a horizontal line, objects coming from above might count as "in" while objects from below count as "out" (or vice versa, depending on line orientation). The triggering anchor determines which point on the bounding box must cross the line for the crossing to be counted - using CENTER ensures the object is substantially across the line before counting.
Common Use Cases¶
- People Counting: Count people entering and exiting buildings, stores, or events (e.g., count visitors entering store, track people entering/exiting building, monitor event attendance), enabling entry/exit counting workflows
- Traffic Analysis: Count vehicles passing through intersections or road segments (e.g., count vehicles crossing intersection, track traffic flow in specific directions, monitor vehicle passage at checkpoints), enabling traffic flow analysis workflows
- Retail Analytics: Track customer movement and foot traffic in retail spaces (e.g., count customers entering store sections, track movement between departments, monitor shopping flow patterns), enabling retail foot traffic analytics workflows
- Security Monitoring: Monitor entry and exit at secure areas or checkpoints (e.g., track entries to restricted areas, count people at access points, monitor checkpoint crossings), enabling security access monitoring workflows
- Occupancy Management: Track occupancy changes by counting objects entering and leaving spaces (e.g., count entries/exits to manage room capacity, track vehicle arrivals/departures in parking, monitor space occupancy changes), enabling occupancy tracking workflows
- Wildlife Monitoring: Count animals crossing defined paths or boundaries (e.g., track animal migration patterns, count wildlife crossing roads, monitor animal movement in habitats), enabling wildlife behavior analysis workflows
Connecting to Other Blocks¶
This block receives tracked detections and video metadata, and produces count_in and count_out values:
- After Byte Tracker blocks to count tracked objects crossing lines (e.g., count tracked people crossing line, track vehicle crossings with consistent IDs, monitor tracked object movements), enabling tracking-to-counting workflows
- After object detection or instance segmentation blocks with tracking enabled to count detected objects (e.g., count detected vehicles, track people crossings, monitor object movements), enabling detection-to-counting workflows
- Before visualization blocks to display line counter information (e.g., visualize line and counts, display crossing statistics, show counting results), enabling counting visualization workflows
- Before data storage blocks to record counting data (e.g., log entry/exit counts, store traffic statistics, record occupancy metrics), enabling counting data logging workflows
- Before notification blocks to alert on count thresholds or events (e.g., alert when count exceeds limit, notify on occupancy changes, trigger actions based on counts), enabling count-based notification workflows
- Before analysis blocks to process counting metrics (e.g., analyze traffic patterns, process occupancy data, work with counting statistics), enabling counting analysis workflows
Requirements¶
This block requires tracked detections with tracker_id information (detections must come from a tracking block like Byte Tracker). The line must be defined as a list of exactly 2 points, where each point is a list or tuple of exactly 2 coordinates (x, y). The block requires video metadata with video_identifier to maintain separate counting state for different videos. The block maintains persistent counting state across frames for each video, so it should be used in video workflows where frames are processed sequentially. For accurate counting, detections should be provided consistently across frames with valid tracker IDs.
Type identifier¶
Use the following identifier in step "type" field: roboflow_core/line_counter@v1to add the block as
as step in your workflow.
Properties¶
| Name | Type | Description | Refs |
|---|---|---|---|
name |
str |
Enter a unique identifier for this step.. | ❌ |
line_segment |
List[Any] |
Line segment defined by exactly two points, each with [x, y] coordinates. Objects crossing from one side count as 'in', objects crossing from the other side count as 'out'. Example: [[0, 100], [500, 100]] creates a horizontal line at y=100. Crossing direction depends on which side objects approach from.. | ✅ |
triggering_anchor |
str |
Point on the bounding box that must cross the line for counting. Options: CENTER (default), BOTTOM_CENTER, TOP_CENTER, CENTER_LEFT, CENTER_RIGHT, etc. CENTER ensures the object is substantially across the line before counting, reducing false positives from objects near but not fully crossing the line.. | ✅ |
The Refs column marks possibility to parametrise the property with dynamic values available
in workflow runtime. See Bindings for more info.
Available Connections¶
Compatible Blocks
Check what blocks you can connect to Line Counter in version v1.
- inputs:
Detections Stitch,Detections Stabilizer,Roboflow Dataset Upload,Object Detection Model,Email Notification,Google Gemma API,Mask Edge Snap,OCR Model,Google Vision OCR,Instance Segmentation Model,Google Gemini,EasyOCR,Detections Combine,Object Detection Model,SAM2 Video Tracker,Detection Event Log,Byte Tracker,ByteTrack Tracker,Bounding Rectangle,OpenAI,Anthropic Claude,Byte Tracker,Detections Consensus,Detections Classes Replacement,Time in Zone,Webhook Sink,Stitch OCR Detections,Object Detection Model,Qwen 3.5 API,Camera Focus,Roboflow Custom Metadata,YOLO-World Model,OpenAI,Detection Offset,Buffer,SAM 3,Instance Segmentation Model,Single-Label Classification Model,Detections List Roll-Up,Size Measurement,VLM As Classifier,Template Matching,Mask Area Measurement,Qwen 3.6 API,SORT Tracker,Florence-2 Model,Detections Transformation,Instance Segmentation Model,CogVLM,Florence-2 Model,Path Deviation,Time in Zone,GLM-OCR,OC-SORT Tracker,S3 Sink,Path Deviation,SAM 3,Seg Preview,Twilio SMS Notification,Model Monitoring Inference Aggregator,Detections Filter,Google Gemini,Local File Sink,Roboflow Dataset Upload,Dynamic Zone,Clip Comparison,Line Counter,Twilio SMS/MMS Notification,Time in Zone,Motion Detection,Dimension Collapse,Anthropic Claude,Clip Comparison,CSV Formatter,VLM As Detector,LMM,Detections Merge,Perspective Correction,Overlap Filter,Anthropic Claude,Velocity,Multi-Label Classification Model,Byte Tracker,SAM 3,Roboflow Vision Events,VLM As Detector,OpenAI,Google Gemini,Dynamic Crop,Per-Class Confidence Filter,Moondream2,Qwen3.5-VL,Segment Anything 2 Model,LMM For Classification,OpenAI,Keypoint Detection Model,MoonshotAI Kimi,Llama 3.2 Vision,Email Notification,PTZ Tracking (ONVIF),Slack Notification,Stitch OCR Detections - outputs:
Line Counter Visualization,Mask Edge Snap,Image Slicer,Instance Segmentation Model,Color Visualization,Ellipse Visualization,ByteTrack Tracker,Polygon Visualization,Byte Tracker,Detections Consensus,Detections Classes Replacement,Webhook Sink,Trace Visualization,Object Detection Model,Stitch OCR Detections,Image Threshold,Heatmap Visualization,SORT Tracker,Halo Visualization,Dot Visualization,Twilio SMS Notification,Dynamic Zone,Pixelate Visualization,Twilio SMS/MMS Notification,Motion Detection,Blur Visualization,Background Subtraction,Text Display,Perspective Correction,Anthropic Claude,Bounding Box Visualization,Stability AI Inpainting,Polygon Visualization,Roboflow Vision Events,Label Visualization,Grid Visualization,Triangle Visualization,Halo Visualization,Circle Visualization,Mask Visualization,Dominant Color,Email Notification,Slack Notification,Detections Stabilizer,Object Detection Model,Stability AI Outpainting,Email Notification,Identify Outliers,Image Preprocessing,Object Detection Model,SAM2 Video Tracker,Byte Tracker,Anthropic Claude,Detection Offset,Instance Segmentation Model,Stitch Images,SIFT Comparison,Morphological Transformation,Instance Segmentation Model,Crop Visualization,OC-SORT Tracker,Icon Visualization,Image Contours,Keypoint Detection Model,Reference Path Visualization,Anthropic Claude,Pixel Color Count,Identify Changes,Classification Label Visualization,Byte Tracker,Absolute Static Crop,Image Blur,Image Slicer,Corner Visualization,Keypoint Detection Model,Keypoint Visualization,QR Code Generator,Morphological Transformation,Keypoint Detection Model,PTZ Tracking (ONVIF),Stitch OCR Detections,SIFT Comparison
Input and Output Bindings¶
The available connections depend on its binding kinds. Check what binding kinds
Line Counter in version v1 has.
Bindings
-
input
metadata(video_metadata): Video metadata containing video_identifier to maintain separate counting state for different videos. Required for persistent counting across frames..detections(Union[object_detection_prediction,instance_segmentation_prediction]): Tracked object detection or instance segmentation predictions. Must include tracker_id information from a tracking block. Objects are counted when their triggering anchor point crosses the line segment..line_segment(list_of_values): Line segment defined by exactly two points, each with [x, y] coordinates. Objects crossing from one side count as 'in', objects crossing from the other side count as 'out'. Example: [[0, 100], [500, 100]] creates a horizontal line at y=100. Crossing direction depends on which side objects approach from..triggering_anchor(string): Point on the bounding box that must cross the line for counting. Options: CENTER (default), BOTTOM_CENTER, TOP_CENTER, CENTER_LEFT, CENTER_RIGHT, etc. CENTER ensures the object is substantially across the line before counting, reducing false positives from objects near but not fully crossing the line..
-
output
Example JSON definition of step Line Counter in version v1
{
"name": "<your_step_name_here>",
"type": "roboflow_core/line_counter@v1",
"metadata": "<block_does_not_provide_example>",
"detections": "$steps.object_detection_model.predictions",
"line_segment": [
[
0,
50
],
[
500,
50
]
],
"triggering_anchor": "CENTER"
}