Electrical & Computer Engineering Department Theses and Dissertations

Permanent URI for this collection

Information about the Purdue School of Engineering and Technology Graduate Degree Programs available at IUPUI can be found at: http://www.engr.iupui.edu/academics.shtml

Browse

Recent Submissions

Now showing 1 - 10 of 222
  • Item
    Development of Automated Fault Recovery Controls for Plug-Flow Biomass Reactors
    (2024-05) Jacob, Mariam; Schubert, Peter J.; Li, Lingxi; King, Brian S.
    The demand for sustainable and renewable energy sources has prompted significant research and development efforts in the field of biomass gasification. Biomass gasification technology holds significant promise for sustainable energy production, offering a renewable alternative to fossil fuels while mitigating environmental impact. This thesis presents a detailed study on the design, development, and implementation of a Plug-Flow Reactor Biomass Gasifier integrated with an Automated Auger Jam Detection System and a Blower Algorithm to maintain constant reactor pressure by varying blower speed with respect to changes in reactor pressure. The system is based on indirectly- heated pyrolytic gasification technology and is developed using Simulink™. The proposed gasification system use the principles of pyrolysis and gasification to convert biomass feedstock into syngas efficiently. An innovative plug-flow reactor configuration ensures uniform heat distribution and residence time, optimizing gasification performance and product quality. Additionally, the system incorporates an automated auger jam detection system, which utilizes sensor data to detect and mitigate auger jams in real-time, thereby enhancing operational reliability and efficiency. By monitoring these parameters, the system detects deviations from normal operating conditions indicative of auger jams and initiates corrective actions automatically. The detection algorithm is trained using test cases and validated through detailed testing to ensure accurate and reliable performance. The MATLAB™-based implementation offers flexibility, scalability, and ease of integration with existing gasifier control systems. The graphical user interface (GUI) provides operators with real-time monitoring and visualization of system status, auger performance, and detected jam events. Additionally, the system generates alerts and notifications to inform operators of detected jams, enabling timely intervention and preventive maintenance. To maintain consistent gasification conditions, a blower algorithm is developed to regulate airflow and maintain constant reactor pressure within the gasifier. The blower algorithm dynamically adjusts blower speed based on feedback from differential pressure sensors, ensuring optimal gasification performance under varying operating conditions. The integration of the blower algorithm into the gasification system contributes to stable syngas production and improved process control. The development of the Plug-Flow Reactor Biomass Gasifier, Automated Auger Jam Detection System, and Blower Algorithm is accompanied by rigorous simulation studies and experimental validation. Overall, this thesis contributes to the advancement of biomass gasification technology by presenting a detailed study on a plug flow reactor biomass gasifier with indirectly- heated pyrolytic gasification technology with an Automated Auger Jam Detection System and Blower Algorithm. The findings offer valuable insights for researchers, engineers, policymakers, and industry stakeholders supporting the transition towards cleaner and more renewable energy systems.
  • Item
    Enhancing Precision of object detectors: bridging classification and localization gaps for 2D and 3D models
    (2024-05) Ravi, Niranjan; El-Sharkawy, Mohamed; Rizkalla, Maher E.; Li, Lingxi; King, Brian S.
    Artificial Intelligence (AI) has revolutionized and accelerated significant advancements in various fields such as healthcare, finance, education, agriculture and the development of autonomous vehicles. We are rapidly approaching Level 5 Autonomy due to recent developments in autonomous technology, including self-driving cars, robot navigation, smart traffic monitoring systems, and dynamic routing. This success has been made possible due to Deep Learning technologies and advanced Computer Vision (CV) algorithms. With the help of perception sensors such as Camera, LiDAR and RADAR, CV algorithms enable a self-driving vehicle to interact with the environment and make intelligent decisions. Object detection lays the foundations for various applications, such as collision and obstacle avoidance, lane detection, pedestrian and vehicular safety, and object tracking. Object detection has two significant components: image classification and object localization. In recent years, enhancing the performance of 2D and 3D object detectors has spiked interest in the research community. This research aims to resolve the drawbacks associated with localization loss estimation of 2D and 3D object detectors by addressing the bounding box regression problem, addressing the class imbalance issue affecting the confidence loss estimation, and finally proposing a dynamic cross-model 3D hybrid object detector with enhanced localization and confidence loss estimation. This research aims to address challenges in object detectors through four key contributions. In the first part, we aim to address the problems associated with the image classification component of 2D object detectors. Class imbalance is a common problem associated with supervised training. Common causes are noisy data, a scene with a tiny object surrounded by background pixels, or a dense scene with too many objects. These scenarios can produce many negative samples compared to positive ones, affecting the network learning and reducing the overall performance. We examined these drawbacks and proposed an Enhanced Hard Negative Mining (EHNM) approach, which utilizes anchor boxes with 20% to 50% overlap and positive and negative samples to boost performance. The efficiency of the proposed EHNM was evaluated using Single Shot Multibox Detector (SSD) architecture on the PASCAL VOC dataset, indicating that the detection accuracy of tiny objects increased by 3.9% and 4% and the overall accuracy improved by 0.9%. To address localization loss, our second approach investigates drawbacks associated with existing bounding box regression problems, such as poor convergence and incorrect regression. We analyzed various cases, such as when objects are inclusive of one another, two objects with the same centres, two objects with the same centres and similar aspect ratios. During our analysis, we observed existing intersections over Union (IoU) loss and its variant’s failure to address them. We proposed two new loss functions, Improved Intersection Over Union (IIoU) and Balanced Intersection Over Union (BIoU), to enhance performance and minimize computational efforts. Two variants of the YOLOv5 model, YOLOv5n6 and YOLOv5s, were utilized to demonstrate the superior performance of IIoU on PASCAL VOC and CGMU datasets. With help of ROS and NVIDIA’s devices, inference speed was observed in real-time. Extensive experiments were performed to evaluate the performance of BIoU on object detectors. The evaluation results indicated MASK_RCNN network trained on the COCO dataset, YOLOv5n6 network trained on SKU-110K and YOLOv5x trained on the custom e-scooter dataset demonstrated 3.70% increase on small objects, 6.20% on 55% overlap and 9.03% on 80% overlap. In the earlier parts, we primarily focused on 2D object detectors. Owing to its success, we extended the scope of our research to 3D object detectors in the later parts. The third portion of our research aims to solve bounding box problems associated with 3D rotated objects. Existing axis-aligned loss functions suffer a performance gap if the objects are rotated. We enhanced the earlier proposed IIoU loss by considering two additional parameters: the objects’ Z-axis and rotation angle. These two parameters aid in localizing the object in 3D space. Evaluation was performed on LiDAR and Fusion methods on 3D KITTI and nuScenes datasets. Once we addressed the drawbacks associated with confidence and localization loss, we further explored ways to increase the performance of cross-model 3D object detectors. We discovered from previous studies that perception sensors are volatile to harsh environmental conditions, sunlight, and blurry motion. In the final portion of our research, we propose a hybrid 3D cross-model detection network (MAEGNN) equipped with MaskedAuto Encoders (MAE) and Graph Neural Networks (GNN) along with earlier proposed IIoU and ENHM. The performance evaluation on MAEGNN on the KITTI validation dataset and KITTI test set yielded a detection accuracy of 69.15%, 63.99%, 58.46% and 40.85%, 37.37% on 3D pedestrians with overlap of 50%. This developed hybrid detector overcomes the challenges of localization error and confidence estimation and outperforms many state-of-art 3D object detectors for autonomous platforms.
  • Item
    Autonomous Detection of Nearby Loss of Generation Events for Decentralized Controls
    (2024-05) Dahal, Niraj; Rovnyak, Steven; Li, Lingxi; Dos Santos, Euzeli; Lee, John
    A broad scope of this dissertation is to verify that a nearby loss of generation event in power system can be distinguished from similar remote disturbances by analyzing the resulting local modes of oscillation. An oscillation-based index derived from methods like Fourier transform, sinc filters and resonant filters is devised and experimented in combination with a variant of df/dt index to jointly classify if a loss of generation event is nearby or remote. A phenomenon widely observed during a loss of generation event is the average decrease in the system’s frequency, typically monitored using the df/dt index. Under-frequency load-shedding (UFLS) relays that are based on df/dt are highly likely to trip for nearby frequency events when combined with the oscillation-based index we propose. Nearby in our context refers to geographical distance, which is correlated with electrical distance, and includes buses within about 50-100 miles of the event location.
  • Item
    Size-Adaptive Convolutional Neural Network with Parameterized-Swish Activation for Enhanced Object Detection
    (2024-05) Venkata Krishnan, Yashwanth Raj; Mohamed, El-Sharkawy; King, Brian; Maher , Rizkalla E.
    In computer vision, accurately detecting objects of varying sizes is essential for various applications, such as autonomous vehicle navigation and medical imaging diagnostics. Addressing the variance in object sizes presents a significant challenge requiring advanced computational solutions for reliable object recognition and processing. This research introduces a size-adaptive Convolutional Neural Network (CNN) framework to enhance detection performance across different object sizes. By dynamically adjusting the CNN’s configuration based on the observed distribution of object sizes, the framework employs statistical analysis and algorithmic decision-making to improve detection capabilities. Further innovation is presented through the Parameterized-Swish activation function. Distinguished by its dynamic parameters, this function is designed to better adapt to varying input patterns. It exceeds the performance of traditional activation functions by enabling faster model convergence and increasing detection accuracy, showcasing the effectiveness of adaptive activation functions in enhancing object detection systems. The implementation of this model has led to notable performance improvements: a 11.4% increase in mean Average Precision (mAP) and a 40.63% increase in frames per second (FPS) for small objects, demonstrating enhanced detection speed and accuracy. The model has achieved a 48.42% reduction in training time for medium-sized objects while still improving mAP, indicating significant efficiency gains without compromising precision. Large objects have seen a 16.9% reduction in training time and a 76.04% increase in inference speed, showcasing the model’s ability to expedite processing times substantially. Collectively, these advancements contribute to a more than 12% increase in detection efficiency and accuracy across various scenarios, highlighting the model’s robustness and adaptability in addressing the critical challenge of size variance in object detection.
  • Item
    A Multi-head Attention Approach with Complementary Multimodal Fusion for Vehicle Detection
    (2024-05) Tabassum, Nujhat; El-Sharkawy, Mohamed; King, Brian; Rizkalla, Maher
    The advancement of autonomous vehicle technologies has taken a significant leap with the development of an improved version of the Multimodal Vehicle Detection Network (MVDNet), distinguished by the integration of a multi-head attention layer. This key enhancement significantly refines the network's capability to process and integrate multimodal sensor data, an aspect that becomes crucial in the face of challenging weather conditions. The effectiveness of this upgraded Multi-Head MVDNet is rigorously verified through an extensive dataset acquired from the Oxford Radar Robotcar, demonstrating its enhanced performance capabilities. Notably, in complex environmental conditions, the Multi-Head MVDNet shows a marked superiority in terms of Average Precision (AP) compared to existing models, underscoring its advanced detection capabilities. The transition from the traditional MVDNet to the enhanced Multi-Head Vehicle Detection Network signifies a notable breakthrough in the arena of vehicle detection technologies, with a special emphasis on operation under severe meteorological conditions, such as the obscuring presence of dense fog or the complexities introduced by heavy snowfall. This significant enhancement capitalizes on the foundational principles of the original MVDNet, which skillfully amalgamates the individual strengths of lidar and radar sensors. This is achieved through an intricate and refined process of feature tensor fusion, creating a more robust and comprehensive sensory data interpretation framework. A major innovation introduced in this updated model is the implementation of a multi-head attention layer. This layer serves as a sophisticated replacement for the previously employed self-attention mechanism. Segmenting the attention mechanism into several distinct partitions enhances the network's efficiency and accuracy in processing and interpreting vast arrays of sensor data. An exhaustive series of experimental analyses was undertaken to determine the optimal configuration of this multi-head attention mechanism. These experiments explored various combinations and settings, ultimately identifying a configuration consisting of seven distinct attention heads as the most effective. This setup was found to optimize the balance between computational efficiency and detection accuracy. When tested using the rich radar and lidar datasets from the ORR project, this advanced Multi-Head MVDNet configuration consistently demonstrated its superiority. It not only surpassed the performance of the original MVDNet but also showed marked improvements over models that relied solely on lidar data or the DEF models, especially in terms of vehicular detection accuracy. This enhancement in the MVDNet model, with its focus on multi-head attention, not only represents a significant leap in the field of autonomous vehicle detection but also lays a foundation for future research. It opens new pathways for exploring various attention mechanisms and their potential applicability in scenarios requiring real-time vehicle detection. Furthermore, it accentuates the importance of sophisticated sensor fusion techniques as vital tools in overcoming the challenges posed by adverse environmental conditions, thus paving the way for more resilient and reliable autonomous vehicular technologies.
  • Item
    Integrating Data-driven Control Methods with Motion Planning: A Deep Reinforcement Learning-based Approach
    (2023-12) Prabu, Avinash; Li, Lingxi; Chen, Yaobin; King, Brian; Tian, Renran
    Path-tracking control is an integral part of motion planning in autonomous vehicles, in which the vehicle's lateral and longitudinal positions are controlled by a control system that will provide acceleration and steering angle commands to ensure accurate tracking of longitudinal and lateral movements in reference to a pre-defined trajectory. Extensive research has been conducted to address the growing need for efficient algorithms in this area. In this dissertation, a scenario and machine learning-based data-driven control approach is proposed for a path-tracking controller. Firstly, a Deep Reinforcement Learning model is developed to facilitate the control of longitudinal speed. A Deep Deterministic Policy Gradient algorithm is employed as the primary algorithm in training the reinforcement learning model. The main objective of this model is to maintain a safe distance from a lead vehicle (if present) or track a velocity set by the driver. Secondly, a lateral steering controller is developed using Neural Networks to control the steering angle of the vehicle with the main goal of following a reference trajectory. Then, a path-planning algorithm is developed using a hybrid A* planner. Finally, the longitudinal and lateral control models are coupled together to obtain a complete path-tracking controller that follows a path generated by the hybrid A* algorithm at a wide range of vehicle speeds. The state-of-the-art path-tracking controller is also built using Model Predictive Control and Stanley control to evaluate the performance of the proposed model. The results showed the effectiveness of both proposed models in the same scenario, in terms of velocity error, lateral yaw angle error, and lateral distance error. The results from the simulation show that the developed hybrid A* algorithm has good performance in comparison to the state-of-the-art path planning algorithms.
  • Item
    Wearable Big Data Harnessing with Deep Learning, Edge Computing and Efficiency Optimization
    (2023-12) Zou, Jiadao; Zhang, Qingxue; King, Brian; Christopher, Lauren; Chien, Stanley
    In this dissertation, efforts and innovations are made to advance subtle pattern mining, edge computing, and system efficiency optimization for biomedical applications, thereby advancing precision medicine big data. Brain visual dynamics encode rich functional and biological patterns of the neural system, promising for applications like intention decoding, cognitive load quantization and neural disorder measurement. We here focus on the understanding of the brain visual dynamics for the Amyotrophic lateral sclerosis (ALS) population. We leverage a deep learning framework for automatic feature learning and classification, which can translate the eye Electrooculography (EOG) signal to meaningful words. We then build an edge computing platform on the smart phone, for learning, visualization, and decoded word demonstration, all in real-time. In a further study, we have leveraged deep transfer learning to boost EOG decoding effectiveness. More specifically, the model trained on basic eye movements is leveraged and treated as an additional feature extractor when classifying the signal to the meaningful word, resulting in higher accuracy. Efforts are further made to decoding functional Near-Infrared Spectroscopy (fNIRS) signal, which encodes rich brain dynamics like the cognitive load. We have proposed a novel Multi-view Multi-channel Graph Neural Network (mmGNN). More specifically, we propose to mine the multi-channel fNIRS dynamics with a multi-stage GNN that can effectively extract the channel- specific patterns, propagate patterns among channels, and fuse patterns for high-level abstraction. Further, we boost the learning capability with multi-view learning to mine pertinent patterns in temporal, spectral, time-frequency, and statistical domains. Massive-device systems, like wearable massive-sensor computers and Internet of Things (IoTs), are promising in the era of big data. The crucial challenge is about how to maximize the efficiency under coupling constraints like energy budget, computing, and communication. We propose a deep reinforcement learning framework, with a pattern booster and a learning adaptor. This framework has demonstrated optimally maximizes the energy utilization and computing efficiency on the local massive devices under a one-center fifteen-device circumstance. Our research and findings are expected to greatly advance the intelligent, real-time, and efficient big data harnessing, leveraging deep learning, edge computing, and efficiency optimization.
  • Item
    Object Detection Using Vision Transformed EfficientDet
    (2023-08) Kar, Shreyanil; El-Sharkawy, Mohamed A.; King, Brian S.; Rizkalla, Maher E.
    This research presents a novel approach for object detection by integrating Vision Transformers (ViT) into the EfficientDet architecture. The field of computer vision, encompassing artificial intelligence, focuses on the interpretation and analysis of visual data. Recent advancements in deep learning, particularly convolutional neural networks (CNNs), have significantly improved the accuracy and efficiency of computer vision systems. Object detection, a widely studied application within computer vision, involves the identification and localization of objects in images. The ViT backbone, renowned for its success in image classification and natural language processing tasks, employs self-attention mechanisms to capture global dependencies in input images. However, ViT’s capability to capture fine-grained details and context information is limited. To address this limitation, the integration of ViT into the EfficientDet architecture is proposed. EfficientDet is recognized for its efficiency and accuracy in object detection. By combining the strengths of ViT and EfficientDet, the proposed integration enhances the network’s ability to capture fine-grained details and context information. It leverages ViT’s global dependency modeling alongside EfficientDet’s efficient object detection framework, resulting in highly accurate and efficient performance. Noteworthy object detection frameworks utilized in the industry, such as RetinaNet, EfficientNet, and EfficientDet, primarily employ convolution. Experimental evaluations were conducted using the PASCAL VOC 2007 and 2012 datasets, widely acknowledged benchmarks for object detection. The integrated ViT-EfficientDet model achieved an impressive mean Average Precision (mAP) score of 86.27% when tested on the PASCAL VOC 2007 dataset, demonstrating its superior accuracy. These results underscore the potential of the proposed integration for real-world applications. In conclusion, the research introduces a novel integration of Vision Transformers into the EfficientDet architecture, yielding significant improvements in object detection performance. By combining ViT’s ability to capture global dependencies with EfficientDet’s efficiency and accuracy, the proposed approach offers enhanced object detection capabilities. Future research directions may explore additional datasets and evaluate the performance of the proposed framework across various computer vision tasks.
  • Item
    User Leaving Detection Via MMwave Imaging
    (2023-08) Xu, Jiawei; King, Brian; Li, Tao; Zhang, Qingxue
    The use of smart devices such as smartphones, tablets, and laptops skyrocketed in the last decade. These devices enable ubiquitous applications for entertainment, communication, productivity, and healthcare but also introduce big concern about user privacy and data security. In addition to various authentication techniques, automatic and immediate device locking based on user leaving detection is an indispensable way to secure the devices. Current user leaving detection techniques mainly rely on acoustic ranging and do not work well in environments with multiple moving objects. In this paper, we present mmLock, a system that enables faster and more accurate user leaving detection in dynamic environments. mmLock uses a mmWave FMCW radar to capture the user’s 3D mesh and detects the leaving gesture from the 3D human mesh data with a hybrid PointNet-LSTM model. Based on explainable user point clouds, mmLock is more robust than existing gesture recognition systems which can only identify the raw signal patterns. We implement and evaluate mmLock with a commercial off-the-shelf (COTS) TI mmWave radar in multiple environments and scenarios. We train the PointNet-LSTM model out of over 1 TB mmWave signal data and achieve 100% true-positive rate in most scenarios.
  • Item
    Utilizing Transfer Learning and Multi-Task Learning for Evaluating the Prediction of Chromatin Accessibility in Cancer and Neuron Cell Lines Using Genomic Sequences
    (2023-08) Shorinwa, Toluwanimi; Salama, Paul; Rizkalla, Maher; El-Sharkawy, Mohamed
    The prediction of chromatin accessibility for cancer and neuron cell lines using genomic sequences is quite challenging. Advances in machine learning and deep learning techniques allow such challenges to be addressed. This thesis investigates the use of both the transfer learning and the multi-task learning techniques. In particular, this research demonstrates the potential of transfer learning and multi-task learning in improving the prediction accuracy for twenty-three cancer types in human and neuron cell lines. Three different network architectures are used: the Basset network, the network, and the DeepSEA network. In addition, two transfer learning techniques are also used. In the first technique data relevant to the desired prediction task is not used during the pre-training stage while the second technique includes limited data about the desired prediction task in the pre-training phase. The preferred performance evaluation metric used to evaluate the performance of the models was the AUPRC due to the numerous negative samples. Our results demonstrate an average improvement of 4% of the DeepSEA network in predicting all twenty-three cancer cell line types when using the first technique, a decrease of 0.42% when using the second technique, and an increase of 0.40% when using multi-task learning. Also, it had an average improvement of 3.09% when using the first technique, 1.16% when using the second technique and 4.60% for the multi-task learning when predicting chromatin accessibility for the 14 neuron cell line types. The DanQ network had an average improvement of 1.18% using the first transfer learning technique, the second transfer learning technique showed an average decrease of 1.93% and also, a decrease of 0.90% for the multi-task learning technique when predicting for the different cancer cell line types. When predicting for the different neuron cell line types the DanQ had an average improvement of 1.56% using the first technique, 3.21% when using the second technique, and 5.35% for the multi-task learning techniques. The Basset network showed an average improvement of 2.93% using the first transfer learning technique and an average decrease of 0.02%, and 0.63% when using the second technique and multi-task learning technique respectively. Using the Basset network for prediction of chromatin accessibility in the different neuron types showed an average increase of 2.47%, 3.80% and 5.50% for the first transfer learning technique, second transfer learning technique and the multi-task learning technique respectively. The results show that the best technique for the cancer cell lines prediction is the first transfer learning model as it showed an improvement for all three network types, while the best technique for predicting chromatin accessibility in the neuron cell lines is the multi-task learning technique which showed the highest average improvement among all networks. The DeepSEA network showed the greatest improvement in performance among all techniques when predicting the different cancer cell line types. Also, it showed the greatest improvement when using the first transfer learning technique for predicting chromatin accessibility for neuron cell lines in the brain. The basset network showed the greatest improvement for the multi-task learning technique and the second transfer learning technique when predicting the accessibility for neuron cell lines.