Clean push: Removed heavy files & added only latest snapshot
This commit is contained in:
57
README.md
Normal file
57
README.md
Normal file
@@ -0,0 +1,57 @@
|
||||
# **Traffic Intersection Monitoring System with OpenVINO**
|
||||
|
||||
This project implements a real-time traffic monitoring solution that detects vehicles, pedestrians, and traffic violations at intersections using object detection models optimized with OpenVINO. It features a PyQt5-based dashboard for visualization and control, integrates synthetic data generation using CARLA, and supports enhanced scene understanding through vision-language models.
|
||||
|
||||
## Problem Statement
|
||||
|
||||
The system monitors traffic intersections to identify and track vehicles, pedestrians, and cyclists in real-time. It collects traffic statistics and detects violations such as red-light running and jaywalking. The focus is on efficient deployment at the edge using Intel hardware.
|
||||
|
||||
## Objectives
|
||||
|
||||
- Detect vehicles, pedestrians, and cyclists using object detection
|
||||
- Monitor and record traffic violations in real-time
|
||||
- Display detection results and statistics through a graphical interface
|
||||
- Enable model deployment using OpenVINO for optimized inference
|
||||
- Generate and annotate synthetic traffic data using CARLA
|
||||
- Integrate visual reasoning capabilities through vision-language models
|
||||
|
||||
## Training and Optimization
|
||||
|
||||
1. **Model Training**
|
||||
The YOLOv12 model is trained using PyTorch with labeled image data representing traffic scenes.
|
||||
|
||||
2. **Export Pipeline**
|
||||
The trained model is exported to ONNX format, and then converted to OpenVINO's Intermediate Representation (IR) format.
|
||||
|
||||
3. **Optimization**
|
||||
Post-training quantization is applied to convert the model from FP32 to INT8, improving inference speed while maintaining accuracy.
|
||||
|
||||
4. **Deployment**
|
||||
OpenVINO's InferRequest API is used for asynchronous inference, enabling efficient frame-by-frame processing suitable for real-time applications.
|
||||
|
||||
## Synthetic Data Generation
|
||||
|
||||
CARLA is used to simulate traffic intersections with accurate layouts, signage, and weather variations. It supports:
|
||||
|
||||
- Scene diversity through environmental changes (rain, fog, glare, nighttime)
|
||||
- Simulation of pedestrian and vehicle behaviors (red-light running, jaywalking)
|
||||
- Automatic annotation of bounding boxes and class labels for use with object detection models
|
||||
|
||||
## Vision-Language Integration
|
||||
|
||||
Two models are integrated to enhance scene understanding:
|
||||
|
||||
- **BLIP-2**: Automatically generates text summaries of visual scenes (e.g., “A vehicle is crossing the red light”)
|
||||
- **LLaVA**: Enables question-answering over video frames (e.g., “Why was the pedestrian flagged?”)
|
||||
|
||||
These tools allow human operators to interact with the system more effectively by supporting natural language explanations and queries.
|
||||
|
||||
## PyQt5-Based Dashboard
|
||||
|
||||
The dashboard enables real-time interaction with the monitoring system and includes:
|
||||
|
||||
- Live video feed with overlayed bounding boxes
|
||||
- Detection tags for pedestrians, vehicles, and violators
|
||||
- Violation statistics and traffic flow metrics
|
||||
- Controls for switching between camera sources and simulated environments
|
||||
- High-performance rendering using QPainter for dynamic visual updates
|
||||
Reference in New Issue
Block a user