2025-08-26 13:17:28 -07:00
2025-08-26 13:07:59 -07:00
2025-08-22 13:18:41 +05:30
2025-08-26 13:17:28 -07:00
2025-08-26 13:07:59 -07:00
2025-08-26 13:07:59 -07:00
2025-08-22 13:18:41 +05:30

Traffic Intersection Monitoring System with OpenVINO

This project implements a real-time traffic monitoring solution that detects vehicles, pedestrians, and traffic violations at intersections using object detection models optimized with OpenVINO. It features a PyQt5-based dashboard for visualization and control, integrates synthetic data generation using CARLA, and supports enhanced scene understanding through vision-language models.

Problem Statement

The system monitors traffic intersections to identify and track vehicles, pedestrians, and cyclists in real-time. It collects traffic statistics and detects violations such as red-light running and jaywalking. The focus is on efficient deployment at the edge using Intel hardware.

Objectives

  • Detect vehicles, pedestrians, and cyclists using object detection
  • Monitor and record traffic violations in real-time
  • Display detection results and statistics through a graphical interface
  • Enable model deployment using OpenVINO for optimized inference
  • Generate and annotate synthetic traffic data using CARLA
  • Integrate visual reasoning capabilities through vision-language models

Training and Optimization

  1. Model Training
    The YOLOv12 model is trained using PyTorch with labeled image data representing traffic scenes.

  2. Export Pipeline
    The trained model is exported to ONNX format, and then converted to OpenVINO's Intermediate Representation (IR) format.

  3. Optimization
    Post-training quantization is applied to convert the model from FP32 to INT8, improving inference speed while maintaining accuracy.

  4. Deployment
    OpenVINO's InferRequest API is used for asynchronous inference, enabling efficient frame-by-frame processing suitable for real-time applications.

Synthetic Data Generation

CARLA is used to simulate traffic intersections with accurate layouts, signage, and weather variations. It supports:

  • Scene diversity through environmental changes (rain, fog, glare, nighttime)
  • Simulation of pedestrian and vehicle behaviors (red-light running, jaywalking)
  • Automatic annotation of bounding boxes and class labels for use with object detection models

Vision-Language Integration

Two models are integrated to enhance scene understanding:

  • BLIP-2: Automatically generates text summaries of visual scenes (e.g., “A vehicle is crossing the red light”)
  • LLaVA: Enables question-answering over video frames (e.g., “Why was the pedestrian flagged?”)

These tools allow human operators to interact with the system more effectively by supporting natural language explanations and queries.

PyQt5-Based Dashboard

The dashboard enables real-time interaction with the monitoring system and includes:

  • Live video feed with overlayed bounding boxes
  • Detection tags for pedestrians, vehicles, and violators
  • Violation statistics and traffic flow metrics
  • Controls for switching between camera sources and simulated environments
  • High-performance rendering using QPainter for dynamic visual updates
Description
No description provided
Readme 11 GiB
Languages
TypeScript 48%
JavaScript 41.8%
Python 9.2%
SCSS 0.6%
CUE 0.2%