Enhanced 3D Object Detection and Tracking in Autonomous Vehicles: An Efficient Multi-Modal Deep Fusion Approach

dc.contributor.advisorEl-Sharkawy, Mohamed
dc.contributor.authorKalgaonkar, Priyank B.
dc.contributor.otherKing, Brian S.
dc.contributor.otherRizkalla, Maher E.
dc.contributor.otherAbdallah, Mustafa A.
dc.date.accessioned2024-09-03T12:58:48Z
dc.date.available2024-09-03T12:58:48Z
dc.date.issued2024-08
dc.degree.date2024
dc.degree.disciplineElectrical & Computer Engineeringen
dc.degree.grantorPurdue Universityen
dc.degree.levelPh.D.
dc.descriptionIUPUI
dc.description.abstractThis dissertation delves into a significant challenge for Autonomous Vehicles (AVs): achieving efficient and robust perception under adverse weather and lighting conditions. Systems that rely solely on cameras face difficulties with visibility over long distances, while radar-only systems struggle to recognize features like stop signs, which are crucial for safe navigation in such scenarios. To overcome this limitation, this research introduces a novel deep camera-radar fusion approach using neural networks. This method ensures reliable AV perception regardless of weather or lighting conditions. Cameras, similar to human vision, are adept at capturing rich semantic information, whereas radars can penetrate obstacles like fog and darkness, similar to X-ray vision. The thesis presents NeXtFusion, an innovative and efficient camera-radar fusion network designed specifically for robust AV perception. Building on the efficient single-sensor NeXtDet neural network, NeXtFusion significantly enhances object detection accuracy and tracking. A notable feature of NeXtFusion is its attention module, which refines critical feature representation for object detection, minimizing information loss when processing data from both cameras and radars. Extensive experiments conducted on large-scale datasets such as Argoverse, Microsoft COCO, and nuScenes thoroughly evaluate the capabilities of NeXtDet and NeXtFusion. The results show that NeXtFusion excels in detecting small and distant objects compared to existing methods. Notably, NeXtFusion achieves a state-of-the-art mAP score of 0.473 on the nuScenes validation set, outperforming competitors like OFT by 35.1% and MonoDIS by 9.5%. NeXtFusion's excellence extends beyond mAP scores. It also performs well in other crucial metrics, including mATE (0.449) and mAOE (0.534), highlighting its overall effectiveness in 3D object detection. Visualizations of real-world scenarios from the nuScenes dataset processed by NeXtFusion provide compelling evidence of its capability to handle diverse and challenging environments.
dc.identifier.urihttps://hdl.handle.net/1805/43087
dc.language.isoen_US
dc.subjectArtificial Intelligence
dc.subjectAI
dc.subjectCNN
dc.subjectDNN
dc.subjectNeural Network
dc.subjectObject Detection
dc.subjectSensor Fusion
dc.titleEnhanced 3D Object Detection and Tracking in Autonomous Vehicles: An Efficient Multi-Modal Deep Fusion Approach
dc.typeThesisen
Files
Original bundle
Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
PhD_Thesis.pdf
Size:
10.08 MB
Format:
Adobe Portable Document Format
License bundle
Now showing 1 - 1 of 1
No Thumbnail Available
Name:
license.txt
Size:
2.04 KB
Format:
Item-specific license agreed upon to submission
Description: