- Browse by Subject
Browsing by Subject "LIDAR"
Now showing 1 - 3 of 3
Results Per Page
Sort Options
Item 3-D Scene Reconstruction for Passive Ranging Using Depth from Defocus and Deep Learning(2019-08) Emerson, David R.; Christopher, Lauren A.; Ben Miled, Zina; King, Brian; Salama, PaulDepth estimation is increasingly becoming more important in computer vision. The requirement for autonomous systems to gauge their surroundings is of the utmost importance in order to avoid obstacles, preventing damage to itself and/or other systems or people. Depth measuring/estimation systems that use multiple cameras from multiple views can be expensive and extremely complex. And as these autonomous systems decrease in size and available power, the supporting sensors required to estimate depth must also shrink in size and power consumption. This research will concentrate on a single passive method known as Depth from Defocus (DfD), which uses an in-focus and out-of-focus image to infer the depth of objects in a scene. The major contribution of this research is the introduction of a new Deep Learning (DL) architecture to process the the in-focus and out-of-focus images to produce a depth map for the scene improving both speed and performance over a range of lighting conditions. Compared to the previous state-of-the-art multi-label graph cuts algorithms applied to the synthetically blurred dataset the DfD-Net produced a 34.30% improvement in the average Normalized Root Mean Square Error (NRMSE). Similarly the DfD-Net architecture produced a 76.69% improvement in the average Normalized Mean Absolute Error (NMAE). Only the Structural Similarity Index (SSIM) had a small average decrease of 2.68% when compared to the graph cuts algorithm. This slight reduction in the SSIM value is a result of the SSIM metric penalizing images that appear to be noisy. In some instances the DfD-Net output is mottled, which is interpreted as noise by the SSIM metric. This research introduces two methods of deep learning architecture optimization. The first method employs the use of a variant of the Particle Swarm Optimization (PSO) algorithm to improve the performance of the DfD-Net architecture. The PSO algorithm was able to find a combination of the number of convolutional filters, the size of the filters, the activation layers used, the use of a batch normalization layer between filters and the size of the input image used during training to produce a network architecture that resulted in an average NRMSE that was approximately 6.25% better than the baseline DfD-Net average NRMSE. This optimized architecture also resulted in an average NMAE that was 5.25% better than the baseline DfD-Net average NMAE. Only the SSIM metric did not see a gain in performance, dropping by 0.26% when compared to the baseline DfD-Net average SSIM value. The second method illustrates the use of a Self Organizing Map clustering method to reduce the number convolutional filters in the DfD-Net to reduce the overall run time of the architecture while still retaining the network performance exhibited prior to the reduction. This method produces a reduced DfD-Net architecture that has a run time decrease of between 14.91% and 44.85% depending on the hardware architecture that is running the network. The final reduced DfD-Net resulted in a network architecture that had an overall decrease in the average NRMSE value of approximately 3.4% when compared to the baseline, unaltered DfD-Net, mean NRMSE value. The NMAE and the SSIM results for the reduced architecture were 0.65% and 0.13% below the baseline results respectively. This illustrates that reducing the network architecture complexity does not necessarily reduce the reduction in performance. Finally, this research introduced a new, real world dataset that was captured using a camera and a voltage controlled microfluidic lens to capture the visual data and a 2-D scanning LIDAR to capture the ground truth data. The visual data consists of images captured at seven different exposure times and 17 discrete voltage steps per exposure time. The objects in this dataset were divided into four repeating scene patterns in which the same surfaces were used. These scenes were located between 1.5 and 2.5 meters from the camera and LIDAR. This was done so any of the deep learning algorithms tested would see the same texture at multiple depths and multiple blurs. The DfD-Net architecture was employed in two separate tests using the real world dataset. The first test was the synthetic blurring of the real world dataset and assessing the performance of the DfD-Net trained on the Middlebury dataset. The results of the real world dataset for the scenes that were between 1.5 and 2.2 meters from the camera the DfD-Net trained on the Middlebury dataset produced an average NRMSE, NMAE and SSIM value that exceeded the test results of the DfD-Net tested on the Middlebury test set. The second test conducted was the training and testing solely on the real world dataset. Analysis of the camera and lens behavior led to an optimal lens voltage step configuration of 141 and 129. Using this configuration, training the DfD-Net resulted in an average NRMSE, NMAE and SSIM of 0.0660, 0.0517 and 0.8028 with a standard deviation of 0.0173, 0.0186 and 0.0641 respectively.Item Comparison of Urban Tree Canopy Classification With High Resolution Satellite Imagery and Three Dimensional Data Derived From LIDAR and Stereoscopic Sensors(2008-08-22T13:59:51Z) Baller, Matthew Lee; Wilson, Jeffrey S. (Jeffrey Scott), 1967-; Tedesco, Lenore P.; Li, LinDespite growing recognition as a significant natural resource, methods for accurately estimating urban tree canopy cover extent and change over time are not well-established. This study evaluates new methods and data sources for mapping urban tree canopy cover, assessing the potential for increased accuracy by integrating high-resolution satellite imagery and 3D imagery derived from LIDAR and stereoscopic sensors. The results of urban tree canopy classifications derived from imagery, 3D data, and vegetation index data are compared across multiple urban land use types in the City of Indianapolis, Indiana. Results indicate that incorporation of 3D data and vegetation index data with high resolution satellite imagery does not significantly improve overall classification accuracy. Overall classification accuracies range from 88.34% to 89.66%, with resulting overall Kappa statistics ranging from 75.08% to 78.03%, respectively. Statistically significant differences in accuracy occurred only when high resolution satellite imagery was not included in the classification treatment and only the vegetation index data or 3D data were evaluated. Overall classification accuracy for these treatment methods were 78.33% for both treatments, with resulting overall Kappa statistics of 51.36% and 52.59%.Item Development of infrared reflectance characteristics of surrogate roadside objects(2018-08) Saha, Abir; Chien, Stanley; Li, Lingxi; Li, LinAn important topic in autonomous vehicle related research in recent times is road departure warning (RDW) and road keeping assistance (RKA). RDW or RKA should be able to recognize and avoid roadside objects. Standard tests are needed to evaluate the performance of RDW and RKA feature of cars from different manufacturers. To avoid damage to the cars under test and the test environment during testing, there is a need of soft, durable and reusable surrogate targets representing various real roadside objects such as curb, concrete divider and metal guardrail. These surrogate objects should have representative characteristics of real roadside objects from the point of view of various commonly used object detection sensors on the vehicles such as camera, radar and LIDAR. Transportation Active Safety Institute (TASI) at Indian University-Purdue University Indianapolis (IUPUI) is in the process of developing surrogate concrete divider, curb metal guardrail and grass that should be recognized as real roadside objects by LIDAR sensors, can be crashed without damage to the test vehicle and can be reused even after multiple crashes. The first step is to understand what the representative roadside objects should look like from the point of view of LIDAR units using laser of various wavelengths, and the next step is to design surrogate objects that successfully emulate the properties of the real roadside objects. Reflectance of an object is an important property for LIDAR detection. This thesis describes an approach for the determination of infrared reflectance property of concrete, metal guardrail and grass for different LIDAR view angles. Various samples of each of these roadside objects were evaluated. Based on these measurements, the suggested reflectance of surrogate roadside objects in the common LIDAR wavelength range of 800-1100 nm is specified. Finally, the design of surrogate roadside objects that satisfy these requirements is described, and the infrared reflectance of these surrogate objects are compared to the suggested reflectance bounds for different LIDAR view angles.