Show simple item record

dc.contributor.authorDey, Nolan Simran
dc.date.accessioned2021-10-01 14:53:23 (GMT)
dc.date.available2021-10-01 14:53:23 (GMT)
dc.date.issued2021-10-01
dc.date.submitted2021-09-28
dc.identifier.urihttp://hdl.handle.net/10012/17608
dc.description.abstractThe field of explainable artificial intelligence (XAI) aims to explain the decisions of DNNs. Complete DNN explanations accurately reflect the inner workings of the DNN while interpretable explanations are easy for humans to understand. Developing methods for explaining the representations learned by DNNs that are both complete and interpretable is a grand challenge in the field of XAI. This thesis makes contributions to the field of XAI by proposing and evaluating novel methods for studying DNN representations. During forward propagation, each DNN layer non-linearly transforms the input space in some way that is useful for minimizing a loss function. To understand how DNNs represent their inputs, this work develops methods to examine each DNN layer’s activation space. The first article contributes an unsupervised framework for identifying and interpreting “tuning dimensions” in the activation space of DNN layers. The method consists of fitting a dimensionality reduction model to a layer’s activations, then visualizing points along the axes defined by each of the reduced dimensions. The method correctly identifies the tuning dimensions of a synthetic Gabor filter bank, and those of the first two layers of InceptionV1 trained on ImageNet. The second article builds upon the first article with a simple and greatly improved visualization method that enables studying every layer of AlexNet. Through a quantitative comparison, the article demonstrates that the principal component analysis (PCA) basis for activation space offers more complete and more interpretable explanations than the traditional neuron basis. This thesis provides deep learning researchers with tools to better understand the representations learned by DNNs.en
dc.language.isoenen
dc.publisherUniversity of Waterlooen
dc.subjectdeep neural networksen
dc.subjectmachine learningen
dc.subjectcomputer visionen
dc.subjectexplainable artificial intelligenceen
dc.subjectinterpretable machine learningen
dc.subject.lcshNeural networks (Computer science)en
dc.subject.lcshMachine learningen
dc.subject.lcshComputer visionen
dc.subject.lcshArtificial intelligenceen
dc.titleStudying CNN representations through activation dimensionality reduction and visualizationen
dc.typeMaster Thesisen
dc.pendingfalse
uws-etd.degree.departmentSystems Design Engineeringen
uws-etd.degree.disciplineSystem Design Engineeringen
uws-etd.degree.grantorUniversity of Waterlooen
uws-etd.degreeMaster of Applied Scienceen
uws-etd.embargo.terms0en
uws.contributor.advisorTripp, Bryan
uws.contributor.advisorTaylor, Graham
uws.contributor.affiliation1Faculty of Engineeringen
uws.published.cityWaterlooen
uws.published.countryCanadaen
uws.published.provinceOntarioen
uws.typeOfResourceTexten
uws.peerReviewStatusUnrevieweden
uws.scholarLevelGraduateen


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record


UWSpace

University of Waterloo Library
200 University Avenue West
Waterloo, Ontario, Canada N2L 3G1
519 888 4883

All items in UWSpace are protected by copyright, with all rights reserved.

DSpace software

Service outages