Skip navigation
Please use this identifier to cite or link to this item: http://arks.princeton.edu/ark:/88435/dsp019880vv175
Title: Exploding and Vanishing Gradients in High-Channel Convolutional Neural Networks
Authors: Domsalla, Wesley
Advisors: Hanin, Boris
Department: Operations Research and Financial Engineering
Certificate Program: Applications of Computing Program
Class Year: 2022
Abstract: We analyze the behavior of the moments of the gradient in a randomly initialized convolutional neural network C with ReLU activations. We show that in randomly initialized networks where the variance of the weights is inversely proportional to the number of channels and kernel sizes, it is possible to avoid the Exploding-Vanishing Gradient problem by scaling the number of channels as the network depth increases; however, the growth rate needed to control the weak bounds we derive for the mo- ments of the gradient is too large to be feasible in practice.
URI: http://arks.princeton.edu/ark:/88435/dsp019880vv175
Type of Material: Princeton University Senior Theses
Language: en
Appears in Collections:Operations Research and Financial Engineering, 2000-2024

Files in This Item:
File Description SizeFormat 
DOMSALLA-WESLEY-THESIS.pdf344.53 kBAdobe PDF    Request a copy


Items in Dataspace are protected by copyright, with all rights reserved, unless otherwise indicated.