A real-time, edge-optimized deep learning perception module for autonomous driving systems. Developed under the Digital Egypt Pioneers Initiative (DEPI) – AI & Machine Learning Track (2025).
Autonomous vehicles require low-latency, high-accuracy environmental perception to ensure safe navigation.
This project presents a YOLOv8-based real-time object detection system optimized for edge deployment using ONNX and TensorRT, achieving:
- ⚡ 78.5 ms inference latency
- 🎯 71.1% mAP@50
- 🚘 Multi-class detection for driving-critical objects
- 🧠 Full MLOps monitoring pipeline
Autonomous systems often struggle with:
- Low-light and night driving
- Fog, rain, and weather interference
- Small object detection (traffic lights, distant pedestrians)
- Real-time latency constraints
We built a perception module that balances:
- ✅ Accuracy
- ✅ Speed
- ✅ Edge deployability
- ✅ Safety-critical robustness
We selected YOLOv8 due to:
- Anchor-free detection head
- C2f backbone blocks
- Efficient gradient flow
- Native ONNX & TensorRT export
- Strong small-object detection performance
Live Camera Feed
↓
Frame Preprocessing (Resize 640x640)
↓
YOLOv8 Inference
↓
NMS & Post-processing
↓
Structured Detection Output
- KITTI (Autonomous Driving Focused)
- COCO (Transfer Learning Base)
- OpenImages (Edge Cases & Diversity)
- Multi-dataset fusion
- Class balancing
- YOLO format conversion
- Scene-leakage prevention
- Annotation normalization
- Night simulation
- Fog & rain injection
- Motion blur
- Mosaic augmentation
- Brightness/contrast shifts
| Parameter | Value |
|---|---|
| Model | YOLOv8 |
| Image Size | 512 |
| Batch Size | 16 |
| Optimizer | AdamW |
| LR | 0.0032 |
| Epochs | 80 |
| Mixed Precision | Enabled |
- Baseline
- Augmentation testing
- Hyperparameter tuning (Optuna)
- Image size benchmarking
- Fine-tuning strategy
Tracked with MLflow.
| Metric | Result |
|---|---|
| Inference Latency | 78.5 ms |
| mAP@50 | 71.1% |
| FPS (Edge Optimized) | 35–45 FPS |
| Target Latency | < 200 ms (Exceeded by 2.5× faster) |
- +12.7% accuracy gain after extended tuning
- Strong generalization in urban scenarios
- Remaining challenge: small traffic light detection
- PyTorch (.pt)
- ONNX (.onnx)
- TensorRT Engine (.engine)
- FP16 quantization
- Threaded inference pipeline
- Hardware-aware tuning
- Jetson validation
Built full monitoring pipeline:
- Model versioning (MLflow)
- Dataset versioning (DVC)
- Drift detection
- FPS & latency monitoring
- Automated retraining triggers
- Confidence distribution tracking
├── data/
├── notebooks/
├── src/
│ ├── training/
│ ├── inference/
│ ├── deployment/
│ └── mlops/
├── models/
├── experiments/
├── exports/
│ ├── onnx/
│ ├── tensorrt/
├── demo/
└── README.md
📺 YouTube Demo: https://youtu.be/UQaHwTuYU6g
- Upgrade to YOLOv8-Medium
- Improve small-object detection
- INT8 Quantization
- Multi-sensor fusion (LiDAR + Camera)
- Real vehicle hardware integration
- Python
- YOLOv8 (Ultralytics)
- PyTorch
- OpenCV
- ONNX
- TensorRT
- MLflow
- DVC
- Optuna
Developed as part of Digital Egyption Pioneers Initiative AI Cohort 2025:
- Elsayed Ali Elsayed Aboulila
- Nizar Hossam Hussein
- Abd El-Rahman Ahmed
- Ahmed Ashraf Abbas
- Mohamed Ashraf Mohamed
Supervised by: Dr. Sherif Salem
This project goes beyond model training - it delivers a deployable, monitored, edge-ready perception module designed for safety-critical autonomous systems.
It demonstrates full-stack ML engineering: From data engineering → model optimization → deployment → monitoring.