Pruning Convolution Neural Network (SqueezeNet) for Efficient Hardware Deployment

dc.contributor.advisorEl-Sharkawy, Mohamed
dc.contributor.authorGaikwad, Akash S.
dc.contributor.otherRizkalla, Maher
dc.contributor.otherKing, Brian
dc.date.accessioned2018-12-05T21:36:36Z
dc.date.available2018-12-05T21:36:36Z
dc.date.issued2018-12
dc.degree.date2018en_US
dc.degree.disciplineElectrical & Computer Engineeringen
dc.degree.grantorPurdue Universityen_US
dc.degree.levelM.S.E.C.E.en_US
dc.descriptionIndiana University-Purdue University Indianapolis (IUPUI)en_US
dc.description.abstractIn recent years, deep learning models have become popular in the real-time embedded application, but there are many complexities for hardware deployment because of limited resources such as memory, computational power, and energy. Recent research in the field of deep learning focuses on reducing the model size of the Convolution Neural Network (CNN) by various compression techniques like Architectural compression, Pruning, Quantization, and Encoding (e.g., Huffman encoding). Network pruning is one of the promising technique to solve these problems. This thesis proposes methods to prune the convolution neural network (SqueezeNet) without introducing network sparsity in the pruned model. This thesis proposes three methods to prune the CNN to decrease the model size of CNN without a significant drop in the accuracy of the model. 1: Pruning based on Taylor expansion of change in cost function Delta C. 2: Pruning based on L2 normalization of activation maps. 3: Pruning based on a combination of method 1 and method 2. The proposed methods use various ranking methods to rank the convolution kernels and prune the lower ranked filters afterwards SqueezeNet model is fine-tuned by backpropagation. Transfer learning technique is used to train the SqueezeNet on the CIFAR-10 dataset. Results show that the proposed approach reduces the SqueezeNet model by 72% without a significant drop in the accuracy of the model (optimal pruning efficiency result). Results also show that Pruning based on a combination of Taylor expansion of the cost function and L2 normalization of activation maps achieves better pruning efficiency compared to other individual pruning criteria and most of the pruned kernels are from mid and high-level layers. The Pruned model is deployed on BlueBox 2.0 using RTMaps software and model performance was evaluated.en_US
dc.identifier.urihttps://hdl.handle.net/1805/17923
dc.identifier.urihttp://dx.doi.org/10.7912/C2/2489
dc.language.isoen_USen_US
dc.rightsAttribution 3.0 United States
dc.rights.urihttp://creativecommons.org/licenses/by/3.0/us/
dc.subjectConvolution neural networken_US
dc.subjectCNNen_US
dc.subjectSqueezeNeten_US
dc.subjectPruningen_US
dc.subjectL2 Normalizationen_US
dc.subjectCIFAR-10en_US
dc.subjectTransfer learningen_US
dc.subjectCoarse pruningen_US
dc.subjectS32V234en_US
dc.subjectTaylor expansionen_US
dc.subjectRTMapsen_US
dc.subjectBlueBoxen_US
dc.subjectFine pruningen_US
dc.subjectModel compressionen_US
dc.subjectActivation mapsen_US
dc.titlePruning Convolution Neural Network (SqueezeNet) for Efficient Hardware Deploymenten_US
dc.typeThesisen
Files
Original bundle
Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
Thesis__Pruning_Convolution_Neural_Network__SqueezeNet__for_Efficient_Hardware_Deployment_.pdf
Size:
4.68 MB
Format:
Adobe Portable Document Format
Description:
Thesis on Pruning Convolution Neural Network (SqueezeNet) for Efficient Hardware Deployment
License bundle
Now showing 1 - 1 of 1
No Thumbnail Available
Name:
license.txt
Size:
1.99 KB
Format:
Item-specific license agreed upon to submission
Description: