Red Zone Monitoring Using Computer Vision
Ensuring the safety of workers is crucial in industrial settings. One effective method to enhance safety is by creating a computer vision system to identify “red zones,” where heavy machinery is passed around, and where workers need to be extremely cautious.
This tutorial will guide you through the process of building a red zone detection system using computer vision. By the end of the guide, you will have a working model that can identify and highlight hazardous areas in real-time.
Step 1. Build a Model
First, sign up for Roboflow and create an account.
Next, go to workspaces and create a project. Customize the project name and annotation group to your choice. Make sure to make an object detection project.
Next, add your images. The images I used were found from this Youtube video. You can enter the Youtube URL into Roboflow to create your dataset quickly.
Then, add the classes you want your model to detect.
Next, start annotating your dataset. We recommend getting at least 50 annotated images before training your first model.
Draw the annotations using the bounding box feature in Roboflow. Repeat this step for each image.
Lastly, generate a dataset version of your labeled images. Each version is unique and associated with a trained model so you can iterate on augmentation and data experiments.
Step 2. Create a Red Zone
Using Roboflow's PolygonZone creator, you can create a red zone where workers should be wary of heavy machinery. Save the points in a variable named “zone” for now.
Step 3. Download needed libraries
We have finished all the prerequisites of model building, we can start to code the application.
First, download the needed libraries:
!pip install roboflow pillow supervision numpy inference
Step 4. Import libraries
Next, import the libraries.
import supervision as sv
import cv2
from typing import Union, List, Optional
from inference.core.interfaces.camera.entities import VideoFrame
from inference import InferencePipeline
from google.colab.patches import cv2_imshow
from datetime import datetime
Make sure to add the polygon zone code.
import numpy as np
zone = np.array([[661, 1638],[1005, 2142],[1649, 1822],[1281, 1342],[665, 1638]])
Step 5. Add detection code
COLOR_ANNOTATOR = sv.ColorAnnotator()
LABEL_ANNOTATOR = sv.LabelAnnotator()
def on_prediction(
predictions: Union[dict, List[Optional[dict]]],
video_frame: Union[VideoFrame, List[Optional[VideoFrame]]],
) -> None:
global zone
if not isinstance(predictions, list):
predictions = [predictions]
video_frame = [video_frame]
for prediction, frame in zip(predictions, video_frame):
if prediction is None:
continue
# SOME PROCESSING
image = frame.image
detections = sv.Detections.from_inference(prediction)
annotated_frame = image
annotated_frame = COLOR_ANNOTATOR.annotate(
scene = annotated_frame,
detections = detections
)
annotated_frame = LABEL_ANNOTATOR.annotate(
scene = annotated_frame,
detections = detections,
)
polyzone = sv.PolygonZone(
polygon=zone,
)
label = "Clear"
zone_presence = polyzone.trigger(detections)
for presence in zone_presence:
if presence:
label = "Occupied"
print("NOT SAFE")
zone_annotated = sv.PolygonZoneAnnotator(
zone=polyzone,
color=sv.Color.RED,
thickness=5,
)
annotated_frame = zone_annotated.annotate(
scene=annotated_frame,
label=label
)
start_point = (2450, 0)
end_point = (2900, 300)
cv2.rectangle(annotated_frame, start_point, end_point, (255, 255, 255), cv2.FILLED)
text = label
font = cv2.FONT_HERSHEY_SIMPLEX
font_scale = 3
font_color = (0, 0, 255)
line_type = 5
center_x = (start_point[0] + end_point[0]) // 2
center_y = (start_point[1] + end_point[1]) // 2
# Get the text size
(text_width, text_height), baseline = cv2.getTextSize(text, font, font_scale, line_type)
# Calculate the bottom-left corner of the text to center it
text_x = center_x - text_width // 2
text_y = center_y + text_height // 2
# Add the text
cv2.putText(annotated_frame, text, (text_x, text_y), font, font_scale, font_color, line_type)
cv2.imshow("frame", annotated_frame)
if cv2.waitKey(1) & 0xFF == ord('q'):
break
In this code snippet, we:
- Get the predictions from each frame
- Add different annotations to each frame using Roboflow Supervision
- Add the polygon to the image
- Check to see if any objects (such as the helmet or boots) are inside of the bounding box
- Edits the top right box if any humans traits are detected within the red zone
Step 6. Piece Everything Together
pipeline = InferencePipeline.init(
video_reference="VIDEO_PATH",
model_id="PATH_TO_MODEL",
max_fps = 30,
confidence=0.8,
api_key="YOUR_API_KEY",
on_prediction=on_prediction,
)
pipeline.start()
Make sure you use the same video you sourced for your model.
Now you are able to detect if a worker is safe or not, depending on their position, as well as if they are in the red zone or not.
Conclusion
By following these steps, you can create a robust red zone detection model that enhances worker safety by identifying dangerous areas.
This guide has walked you through the process of setting up a model using Roboflow, defining red zones, and integrating the detection system with real-time video feeds.
For more computer vision projects, check out our other guides and resources.