AI on the Edge with CondenseNeXt: An Efficient Deep Neural Network for Devices with Constrained Computational Resources

dc.contributor.advisorEl-Sharkawy, Mohamed A.
dc.contributor.authorKalgaonkar, Priyank B.
dc.contributor.otherKing, Brian S.
dc.contributor.otherRizkalla, Maher E.
dc.date.accessioned2021-08-10T13:26:32Z
dc.date.available2021-08-10T13:26:32Z
dc.date.issued2021-08
dc.degree.date2021en_US
dc.degree.disciplineElectrical & Computer Engineeringen
dc.degree.grantorPurdue Universityen_US
dc.degree.levelM.S.E.C.E.en_US
dc.descriptionIndiana University-Purdue University Indianapolis (IUPUI)en_US
dc.description.abstractResearch work presented within this thesis propose a neoteric variant of deep convolutional neural network architecture, CondenseNeXt, designed specifically for ARM-based embedded computing platforms with constrained computational resources. CondenseNeXt is an improved version of CondenseNet, the baseline architecture whose roots can be traced back to ResNet. CondeseNeXt replaces group convolutions in CondenseNet with depthwise separable convolutions and introduces group-wise pruning, a model compression technique, to prune (remove) redundant and insignificant elements that either are irrelevant or do not affect performance of the network upon disposition. Cardinality, a new dimension to the existing spatial dimensions, and class-balanced focal loss function, a weighting factor inversely proportional to the number of samples, has been incorporated in order to relieve the harsh effects of pruning, into the design of CondenseNeXt’s algorithm. Furthermore, extensive analyses of this novel CNN architecture was performed on three benchmarking image datasets: CIFAR-10, CIFAR-100 and ImageNet by deploying the trained weight on to an ARM-based embedded computing platform: NXP BlueBox 2.0, for real-time image classification. The outputs are observed in real-time in RTMaps Remote Studio’s console to verify the correctness of classes being predicted. CondenseNeXt achieves state-of-the-art image classification performance on three benchmark datasets including CIFAR-10 (4.79% top-1 error), CIFAR-100 (21.98% top-1 error) and ImageNet (7.91% single model, single crop top-5 error), and up to 59.98% reduction in forward FLOPs compared to CondenseNet. CondenseNeXt can also achieve a final trained model size of 2.9 MB, however at the cost of 2.26% in accuracy loss. Thus, performing image classification on ARM-Based computing platforms without requiring a CUDA enabled GPU support, with outstanding efficiency.en_US
dc.identifier.urihttps://hdl.handle.net/1805/26438
dc.identifier.urihttp://dx.doi.org/10.7912/C2/64
dc.language.isoen_USen_US
dc.subjectCondenseNeXten_US
dc.subjectedgeen_US
dc.subjectefficienten_US
dc.subjectmodelen_US
dc.subjectCIFAR10en_US
dc.subjectCIFAR100en_US
dc.subjectPythonen_US
dc.subjectPyTorchen_US
dc.subjectImage Processingen_US
dc.subjectComputer Visionen_US
dc.subjectDeep Learningen_US
dc.subjectNeural Networksen_US
dc.subjectAIen_US
dc.subjectImage Classificationen_US
dc.subjectMachine Learningen_US
dc.subjectArtificial Intelligenceen_US
dc.subjectCondenseNeten_US
dc.subjectalgorithmen_US
dc.subjectCNNen_US
dc.subjectDNNen_US
dc.subjectANNen_US
dc.subjectNNen_US
dc.subjectedge computingen_US
dc.titleAI on the Edge with CondenseNeXt: An Efficient Deep Neural Network for Devices with Constrained Computational Resourcesen_US
dc.typeThesisen
Files
Original bundle
Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
Priyank Final Masters Thesis 06_01.pdf
Size:
44.97 MB
Format:
Adobe Portable Document Format
Description:
Thesis Document
License bundle
Now showing 1 - 1 of 1
No Thumbnail Available
Name:
license.txt
Size:
1.99 KB
Format:
Item-specific license agreed upon to submission
Description: