A component based knowledge transfer model for deep neural networks: [a thesis submitted to Auckland University of Technology in fulfilment of the requirements for the degree of Doctor of Philosophy (PhD), 2020] / Sreenivas Sremath Tirumala ; supervisors: Ajit Narayanan, Jacqueline Whalley.

This thesis explores the idea that features extracted from deep neural networks (DNNs) through layered weight analysis are knowledge components and are transferable. Among the components extracted from the various layers, middle layer components are shown to constitute knowledge that is mainly respo...

Full description

Saved in:
Bibliographic Details
Main Author: Sremath Tirumala, Sreenivas (Author)
Corporate Author: Auckland University of Technology
Format: Ethesis
Language:English
Subjects:
Online Access:Click here to access this resource online
Description
Summary:This thesis explores the idea that features extracted from deep neural networks (DNNs) through layered weight analysis are knowledge components and are transferable. Among the components extracted from the various layers, middle layer components are shown to constitute knowledge that is mainly responsible for the accuracy of deep architectures including deep autoencoders (DAEs), deep belief networks (DBNs) and DNNs. The proposed component-based transfer of knowledge is shown to be efficient when applied to a variety of benchmark datasets including handwritten character recognition, image recognition, speech analysis, gene expression, as well as hierarchical feature datasets. The importance of hidden layer and its position in the topology of Artificial Neural Networks (ANNs) is under-researched in comparison to the deployment of new architectures, components and learning algorithms. This thesis addresses this imbalance by providing an insight into what actually is learned by a neural network. This is because recent advances in layer-wise training enable us to explore systematically and rigorously the features that expose hidden layer by hidden layer in deep architectures. The key contribution of this research is providing a transferable component model by extracting knowledge components from hidden layers. This thesis also provides an approach to determine the contribution of individual layers, thus providing an insight into the topological constraints that require addressing while designing a transfer learning model. Such transfer learning can mitigate the problem of needing to train each neural network 'from scratch.' This is important since deep learning currently can be slow and require large amounts of processing power. "Warm started" deep learning may open new avenues of research, especially in areas where 'portable' deep architectures can be deployed for decision making.
Author supplied keywords: Deep Learning; Feature Extraction; Knowledge Discovery; Transferable knowledge; Artificial Neural Networks; Deep Nerual Networks.
Physical Description:1 online resource
Bibliography:Includes bibliographical references.
Requests
Request this item Request this AUT item so you can pick it up when you're at the library.
Interlibrary Loan With Interlibrary Loan you can request the item from another library. It's a free service.