Squeeze and Excite Residual Capsule Network for Embedded Edge Devices

dc.contributor.advisorEl-Sharkawy, Mohamed
dc.contributor.authorNaqvi, Sami
dc.contributor.otherKing, Brian
dc.contributor.otherRizkalla, Maher
dc.date.accessioned2022-09-15T11:08:40Z
dc.date.available2022-09-15T11:08:40Z
dc.date.issued2022-08
dc.degree.date2022en_US
dc.degree.disciplineElectrical & Computer Engineeringen
dc.degree.grantorPurdue Universityen_US
dc.degree.levelM.S.E.C.E.en_US
dc.descriptionIndiana University-Purdue University Indianapolis (IUPUI)en_US
dc.description.abstractDuring recent years, the field of computer vision has evolved rapidly. Convolutional Neural Networks (CNNs) have become the chosen default for implementing computer vision tasks. The popularity is based on how the CNNs have successfully performed the well-known computer vision tasks such as image annotation, instance segmentation, and others with promising outcomes. However, CNNs have their caveats and need further research to turn them into reliable machine learning algorithms. The disadvantages of CNNs become more evident as the approach to breaking down an input image becomes apparent. Convolutional neural networks group blobs of pixels to identify objects in a given image. Such a technique makes CNNs incapable of breaking down the input images into sub-parts, which could distinguish the orientation and transformation of objects and their parts. The functions in a CNN are competent at learning only the shift-invariant features of the object in an image. The discussed limitations provides researchers and developers a purpose for further enhancing an effective algorithm for computer vision. The opportunity to improve is explored by several distinct approaches, each tackling a unique set of issues in the convolutional neural network’s architecture. The Capsule Network (CapsNet) which brings an innovative approach to resolve issues pertaining to affine transformations by sharing transformation matrices between the different levels of capsules. While, the Residual Network (ResNet) introduced skip connections which allows deeper networks to be more powerful and solves vanishing gradient problem. The motivation of these fusion of these advantageous ideas of CapsNet and ResNet with Squeeze and Excite (SE) Block from Squeeze and Excite Network, this research work presents SE-Residual Capsule Network (SE-RCN), an efficient neural network model. The proposed model, replaces the traditional convolutional layer of CapsNet with skip connections and SE Block to lower the complexity of the CapsNet. The performance of the model is demonstrated on the well known datasets like MNIST and CIFAR-10 and a substantial reduction in the number of training parameters is observed in comparison to similar neural networks. The proposed SE-RCN produces 6.37 Million parameters with an accuracy of 99.71% on the MNIST dataset and on CIFAR-10 dataset it produces 10.55 Million parameters with 83.86% accuracy.en_US
dc.identifier.urihttps://hdl.handle.net/1805/30003
dc.identifier.urihttp://dx.doi.org/10.7912/C2/3026
dc.language.isoen_USen_US
dc.subjectComputer Visionen_US
dc.subjectConvolutional Neural Networksen_US
dc.subjectCapsule Networken_US
dc.subjectDynamic Routingen_US
dc.titleSqueeze and Excite Residual Capsule Network for Embedded Edge Devicesen_US
dc.typeThesisen
Files
Original bundle
Now showing 1 - 1 of 1
No Thumbnail Available
Name:
Sami_Naqvi_Thesis_Purdue_University (1).pdf
Size:
4.46 MB
Format:
Adobe Portable Document Format
Description:
License bundle
Now showing 1 - 1 of 1
No Thumbnail Available
Name:
license.txt
Size:
1.99 KB
Format:
Item-specific license agreed upon to submission
Description: