This repository showcases the technical aspects of a comprehensive computer vision system developed for autonomous vehicles. The project integrates multiple state-of-the-art computer vision techniques with sensor data fusion to ensure accurate perception, navigation, and decision-making capabilities for autonomous vehicles.
-
Object Detection and Tracking
- Real-time detection and tracking of vehicles, pedestrians, traffic lights, and road signs using advanced deep learning models (e.g., YOLOv5, YOLOP).
- A custom YOLOv5 model trained on traffic sign data and a traffic light classification model were loaded for detecting relevant objects and traffic signals and the signal colors (red, yellow, and green). We fine-tuned the parameters, confidence thresholds and IoU thresholds for optimal detection accuracy. A lightweight convolutional neural network (CNN) is used to classify traffic light colors.
- Optimization to reduce false positives and increase detection range for distant objects.
-
Lane Detection
- Robust lane detection algorithms capable of identifying lanes under varying lighting and weather conditions.
- Dynamic region of interest (ROI) optimization for efficient computation.
- Depth Estimation for Distance Measurement
- The system calculates the distance to detected objects using depth information, providing warnings when objects come within a predefined safety range, thereby enhancing vehicle safety.
- A side panel displays dynamic warnings and vehicle information, including lane curvature, vehicle position, and potential hazards.
- We also subscribed to camera topic on ROS and published proximity warning. The topic published true upon detection of object within the set distance threshold and false if far.
-
Sensor Fusion
- Combining data from multiple sensors, including cameras, LiDAR, and GPS, to enhance perception and navigation accuracy.
- Real-time data synchronization and processing.
-
Path Planning and Control
- Basic motion planning experiments with ROS tools.
- Incorporation of traffic light signals into autonomous vehicle decision-making.
Captured using ZED stereo cameras in .bag format.
Object detection and distance data saved in .csv files for further analysis.
Heatmaps and 3D visualizations created using tools like RViz and Python libraries (e.g., Folium, Matplotlib).
For further technical details, access to the codebase, or collaboration inquiries, please reach out to us at:
Email: [email protected]
LinkedIn: Fatima Saud Linkedin