Project Report
Project Report
Project Report
on
Vision Drive: Smart Object Detection for Autonomous
Vehicles
submitted as partial fulfillment for the award of
BACHELOR OF TECHNOLOGY
DEGREE
SESSION 2024-25
in
COMPUTER SCIENCE & ENGINEERING
By
Yash Tyagi (2102311530059)
Harsh Tyagi (2102311530027)
Varsha (2102311530056)
Sakshi Bhardwaj (2102311530045)
We hereby declare that this submission is our own work and that, to the best of our
knowledge and belief, it contains no material previously published or written by
another person nor material which to a substantial extent has been accepted for
the award of any other degree or diploma of the university or other institute of
higher learning, except where due acknowledgment has been made in the text.
Signature :
Name :
Roll No. :
Date :
Signature :
Name :
Roll No. :
Date :
Signature :
Name :
Roll No. :
Date :
Signature :
Name :
Roll No. :
I.
Date :
CERTIFICATE
This is to certify that Project Report entitled – “ vision drive : smart object
detection for autonomous vehicles” which is submitted by yash
tyagi(2102311530059), harsh tyagi(2102311530027) ,varsha(2102311530056), sakshi
bhardwaj(2102311530045),in partial fulfillment of the requirement for the award of degree
B. Tech. in Department of CSE, of Dr. A.P.J. Abdul Kalam Technical University, U.P.,
Lucknow., is a record of the candidate own work carried out by him/her under my/our
supervision. The matter embodied in this Project report is original and has not been
submitted for the award of any other degree.
Date:
I.
ACKNOWLEDGEMENT
It gives us a great sense of pleasure to present the report of the B. Tech Project
undertaken during B.Tech Final Year. We owe special debt of gratitude to our guide
Prof. Lav Kumar Dixit , Department of CSE, R.D. Engineering College, Ghaziabad for
his constant support and guidance throughout the course of our work. His sincerity,
thoroughness and perseverance have been a constant source of inspiration for us. It is
only his cognizant efforts that our endeavours have seen light of the day.
We express our sincere gratitude to Prof. Lav Kumar Dixit, HoD, Department of CSE,
R.D. Engineering College and Research Centre, Ghaziabad, for his stimulating
guidance, continuous encouragement and supervision during the development of the
project.
We are extremely thankful to Prof. Mohd. Vakil, Dean Academics, R.D. Engineering
College, Ghaziabad, for his full support and assistance during the development of the
project.
We also do not like to miss the opportunity to acknowledge the contribution of all
faculty members of the department for their kind assistance and cooperation during
the development of our project. Last but not the least, we acknowledge our friends for
their contribution in the completion of the project.
Signat : Signature :
ure
:yash tyagi Name :harsh tyagi
Name
Roll : 2102311530059 Roll No. : 2102311530027
No.
Date : Date :
Signat : Signat
ure ure
:varsha : sakshi
Name Name bhardwaj
Roll :2102311530056 Roll :210231153
No. No. 0045
Date : Date :
ABSTRACT
Autonomous vehicles rely on sophisticated object detection technology to facilitate efficient and
safe driving. This technology combines sensor fusion methods, machine learning, and computer
vision techniques to detect and track objects such as pedestrians, obstacles, and other vehicles in
real-time. This project examines state-of-the-art object detection models, including YOLO, Faster
R-CNN, and SSD, and their contributions to enhancing the perception capabilities of autonomous
vehicles.
Despite advancements, challenges such as occlusion, illumination changes, and computational costs
persist. The project explores multi-sensor fusion techniques that integrate data from cameras,
LiDAR, and radar to improve detection accuracy and robustness. Experimental results demonstrate
high detection accuracy, with YOLO-based models achieving over 90% mean Average Precision
(mAP) on benchmark datasets like KITTI and COCO. The system's deployment on edge platforms
ensures real-time performance, making it suitable for autonomous driving applications.
CHAPTER 1 INTRODUCTION 1
1.1 Autonomous Vehicles: Technological and Societal Transformation
1.2 Object Detection: Theoretical Foundations
1.2.1 Fundamental Computer Vision Techniques
1.2.2 Sensor-Specific Detection Challenges
1.3 Deep Learning Architectures for Autonomous Driving
1.3.1 Two-Stage Detectors
1.3.2 Single-Stage Detectors
1.3.3 Emerging Transformer Models
1.4 VisionDrive Project: System-Level Innovation
1.4.1 Novel Contributions
1.4.2 Real-World Validation
3.1 Introduction
3.2 System Overview
3.3 Module Description
3.3.1 Data Acquisition Module
3.3.2 Object Detection Module
3.3.3 Decision Support Module
3.4 Entity Relationship (ER) Diagram
3.5 Data Flow Diagram (DFD)
3.6 Flow Chart
3.7 Algorithm
3.8 Key Equations
3.9 Dataset Description
3.10 UML Diagrams
3.11 Feasibility Study
3.12 Hardware Requirements
3.13 Software Requirements
3.14 Implementation Plan
3.15 Summary
4.5 Conclusion
REFERENCES 6
● Market Projections:
1.2 Object Detection: Theoretical Foundations
● Radar Limitations:
1.Angular resolution vs. detection range tradeoff (Fig 1.3)
2.Doppler ambiguity in urban environments
1.3 Deep Learning Architectures for Autonomous Driving
● TEST SCENARIOS :
1.Urban (SF): 92.1% mAP
2.Highway (I-80): 94.3% mAP
3.Adverse Weather:
3.1Rain: 87.5% mAP
3.2 Fog: 83.2% mAP
● FAILURE CASE ANALYSIS :
1 Occlusion Handling:
1.1Baseline: 68% recall VisionDrive: 81% recall
2 Sensor Failure Modes:
2.2 Camera glare recovery time: 2.3s 1.1s with LiDAR backup
** Visual Appendices **
2.8 Conclusion
3.1 Introduction
This chapter presents the comprehensive methodology for the
**VisionDrive** smart object detection system for autonomous
vehicles. The proposed solution integrates **deep learning
algorithms, multi-sensor fusion, and edge computing** to
achieve real-time, robust object detection. The methodology is
structured to cover all critical aspects from system architecture
to implementation details.
3.2 System Overview
The VisionDrive system comprises three core modules:
1. Data Acquisition Module
2. Object Detection Module
3. Decision Support Module
3.3 Module Description
3.3.1 Data Acquisition Module
● Input Sources:
- Cameras (RGB, stereo, thermal)
- LiDAR (64-channel)
- Radar (77GHz)
● Synchronization :
- Hardware-level time synchronization
- Kalman filtering for temporal alignment
Level 1 DFD:
+---------------+
| Sensor Inputs | --> [Data Preprocessing] --> [Object Detection]
+---------------+ | |
v v
[Calibration Module] [Fusion Engine]
|
v
[Decision Support System]
3.15 Summary
This chapter presented a detailed methodology covering:
- System architecture and modules
- Technical diagrams and algorithms
- Dataset and hardware specifications
- Comprehensive feasibility analysis
time object detection. Below are the key components, screenshots, and
● Hardware :
● Software:
Data from LiDAR (point clouds), camera (RGB images), and radar
(velocity/range) are synchronized using **Kalman filters** and processed
via:
● Metrics :
|--------------------------|-------------------|---------------------------|
| YOLOv4 | 90.1 | 45 |
1. Camera-Only Detection
[Camera Output](project_imp.png)
2. LiDAR-Camera Fusion
[Fusion Output](project_1.png)
5.1 Conclusion
The rapid advancements in autonomous vehicle (AV) technology have made object
detection a critical component for ensuring safe and efficient navigation. This
research explored the state-of-the-art object detection models, including **YOLO
(You Only Look Once), Faster R-CNN, and SSD (Single Shot MultiBox Detector)**,
and their applicability in autonomous driving scenarios. These models leverage deep
learning techniques, particularly **Convolutional Neural Networks (CNNs) and
transformer-based architectures**, to achieve high accuracy in real-time object
detection.
This research underscores the critical role of deep learning and sensor fusion in
advancing autonomous vehicle perception systems. While current models like
YOLOv4, Faster R-CNN, and SSD provide a strong foundation, future innovations in
transformer architectures, edge AI, and adaptive learning will drive the next
generation of AV technology. The proposed VisionDrive framework demonstrates the
feasibility of real-time, robust object detection, paving the way for fully autonomous
and safe transportation systems.
1. Joseph Redmon, Santosh Divvala, Ross Girshick, Ali Farhadi, "You Only
Look Once: Unified, Real-Time Object Detection," IEEE Conference on
Computer Vision and Pattern Recognition (CVPR), pp. 779-788, 2016.
3. Shaoqing Ren, Kaiming He, Ross Girshick, Jian Sun, "Faster R-CNN:
Towards Real-Time Object Detection with Region Proposal Networks,"
Advances in Neural Information Processing Systems (NIPS) , pp. 91-99, 2015.
8. Siyuan Liang, Hao Wu, "Edge YOLO: Real-Time Intelligent Object Detection
System Based on Edge-Cloud Cooperation in Autonomous Vehicles," arXiv
preprint arXiv:2205.15472, May 2022.