Analysis of Latent Space Representations for Object Detection

dc.contributor.advisorChristopher, Lauren
dc.contributor.authorDale, Ashley Susan
dc.contributor.otherKing, Brian
dc.contributor.otherSalama, Paul
dc.contributor.otherRizkalla, Maher
dc.date.accessioned2024-09-03T12:33:25Z
dc.date.available2024-09-03T12:33:25Z
dc.date.issued2024-08
dc.degree.date2024
dc.degree.disciplineElectrical & Computer Engineeringen
dc.degree.grantorPurdue University
dc.degree.levelPh.D.
dc.descriptionIndiana University-Purdue University Indianapolis (IUPUI)en
dc.description.abstractDeep Neural Networks (DNNs) successfully perform object detection tasks, and the Con- volutional Neural Network (CNN) backbone is a commonly used feature extractor before secondary tasks such as detection, classification, or segmentation. In a DNN model, the relationship between the features learned by the model from the training data and the features leveraged by the model during test and deployment has motivated the area of feature interpretability studies. The work presented here applies equally to white-box and black-box models and to any DNN architecture. The metrics developed do not require any information beyond the feature vector generated by the feature extraction backbone. These methods are therefore the first methods capable of estimating black-box model robustness in terms of latent space complexity and the first methods capable of examining feature representations in the latent space of black box models. This work contributes the following four novel methodologies and results. First, a method for quantifying the invariance and/or equivariance of a model using the training data shows that the representation of a feature in the model impacts model performance. Second, a method for quantifying an observed domain gap in a dataset using the latent feature vectors of an object detection model is paired with pixel-level augmentation techniques to close the gap between real and synthetic data. This results in an improvement in the model’s F1 score on a test set of outliers from 0.5 to 0.9. Third, a method for visualizing and quantifying similarities of the latent manifolds of two black-box models is used to correlate similar feature representation with increase success in the transferability of gradient-based attacks. Finally, a method for examining the global complexity of decision boundaries in black-box models is presented, where more complex decision boundaries are shown to correlate with increased model robustness to gradient-based and random attacks.
dc.identifier.urihttps://hdl.handle.net/1805/43077
dc.language.isoen_US
dc.rightsAttribution 4.0 Internationalen
dc.rights.urihttps://creativecommons.org/licenses/by/4.0
dc.subjectComputer Vision
dc.subjectArtificial Intelligence
dc.subjectAdversarial machine learning
dc.subjectPattern Recognition
dc.subjectImage processing
dc.subjectKnowledge representation and reasoning
dc.titleAnalysis of Latent Space Representations for Object Detection
dc.typeThesisen
Files
Original bundle
Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
ANALYSIS_OF_LATENT_SPACE_REPRESENTATION_FOR_OBJECT_DETECTION.pdf
Size:
44.94 MB
Format:
Adobe Portable Document Format
License bundle
Now showing 1 - 1 of 1
No Thumbnail Available
Name:
license.txt
Size:
2.04 KB
Format:
Item-specific license agreed upon to submission
Description: