Graduation Semester and Year
2018
Language
English
Document Type
Thesis
Degree Name
Master of Science in Computer Science
Department
Computer Science and Engineering
First Advisor
Vassilis Athitsos
Abstract
Deep Neural Network have become very popular for computer vision application in recent years. At the same time, it remains important to understand the different implementation choices that need to be made when designing a neural network and to thoroughly investigate existing and novel alternatives for those choices. One of those choices is the activation function. The ReLU activation function is a widely used activation function. It discards all the values below zero and keeps the ones greater than zero. Variations such as Leaky ReLU and Parametric ReLU do not discard values, so that gradiants are nonzero for the entire input range. However, one or both scaling parameters are implicitly or explicitly hardcoded. We are proposing a new variation of ReLU, that we call Double-Weighted Rectifier Liner Unit (DWReLU), in which both scaling parameters are trainable. In our experiment, on popular bench mark datasets (MINIST and CIFAR-10), the praposed activation function leads to better accuracy most of the time, compared to other activation function.
Keywords
Deep neural networks, Activation functions, Machine learning
Disciplines
Computer Sciences | Physical Sciences and Mathematics
License
This work is licensed under a Creative Commons Attribution-NonCommercial-Share Alike 4.0 International License.
Recommended Citation
Trivedi, Bhaskar Chandra, "DWRELU : DOUBLE WEIGHTED RECTIFIER LINEAR UNIT AN ACTIVATION FUNCTION WITH TRAINABLE SCALING PARAMETER" (2018). Computer Science and Engineering Theses. 400.
https://mavmatrix.uta.edu/cse_theses/400
Comments
Degree granted by The University of Texas at Arlington