Please use this identifier to cite or link to this item:
http://arks.princeton.edu/ark:/88435/dsp019880vv175
Title: | Exploding and Vanishing Gradients in High-Channel Convolutional Neural Networks |
Authors: | Domsalla, Wesley |
Advisors: | Hanin, Boris |
Department: | Operations Research and Financial Engineering |
Certificate Program: | Applications of Computing Program |
Class Year: | 2022 |
Abstract: | We analyze the behavior of the moments of the gradient in a randomly initialized convolutional neural network C with ReLU activations. We show that in randomly initialized networks where the variance of the weights is inversely proportional to the number of channels and kernel sizes, it is possible to avoid the Exploding-Vanishing Gradient problem by scaling the number of channels as the network depth increases; however, the growth rate needed to control the weak bounds we derive for the mo- ments of the gradient is too large to be feasible in practice. |
URI: | http://arks.princeton.edu/ark:/88435/dsp019880vv175 |
Type of Material: | Princeton University Senior Theses |
Language: | en |
Appears in Collections: | Operations Research and Financial Engineering, 2000-2024 |
Files in This Item:
File | Description | Size | Format | |
---|---|---|---|---|
DOMSALLA-WESLEY-THESIS.pdf | 344.53 kB | Adobe PDF | Request a copy |
Items in Dataspace are protected by copyright, with all rights reserved, unless otherwise indicated.