Date of Award
2016
Document Type
Thesis
Degree Name
M.S. in Engineering Science
Department
Computer and Information Science
First Advisor
Dawn Wilkins
Second Advisor
Byunghyun Jang
Third Advisor
Yixin Chen
Relational Format
dissertation/thesis
Abstract
Deep neural networks (DNNs), and artificial neural networks (ANNs) in general, have recently received a great amount of attention from both the media and the machine learning community at large. DNNs have been used to produce world-class results in a variety of domains, including image recognition, speech recognition, sequence modeling, and natural language processing. Many of most exciting recent deep neural network studies have made improvements by hardcoding less about the network and giving the neural network more control over its own parameters, allowing flexibility and control within the network. Although much research has been done to introduce trainable hyperparameters into transformation layers (GRU [7], LSTM [13], etc), the introduction of hyperparameters into the activation layers have been largely ignored. This paper serves several purposes: to (1) equip the reader with the background knowledge, including theory and best practices for DNNs, which help contextualize the contributions of this paper, (2) to describe and verify the effectiveness of current techniques in the literature that utilize hyperparameters in the activation layer, and (3) to introduce some new activation layers that introduce hyperparameters into the model, including activation pools (APs) and parametric activation pools (PAPs), and study the effectiveness of these new constructs on popular image recognition datasets.
Recommended Citation
Mcleod, Clay Lafayette, "The Effect Of Hyperparameters In The Activation Layers Of Deep Neural Networks" (2016). Electronic Theses and Dissertations. 451.
https://egrove.olemiss.edu/etd/451
Concentration/Emphasis
Emphasis: Computer Science