Skip Nav Destination
Close Modal
Update search
NARROW
Format
Journal
Date
Availability
1-1 of 1
Hanie Sedghi
Close
Follow your search
Access your saved searches in your account
Would you like to receive an alert when new items match your search?
Sort by
Journal Articles
On the Effect of the Activation Function on the Distribution of Hidden Nodes in a Deep Network
UnavailablePublisher: Journals Gateway
Neural Computation (2019) 31 (12): 2562–2580.
Published: 01 December 2019
FIGURES
Abstract
View articletitled, On the Effect of the Activation Function on the Distribution of Hidden Nodes in a Deep Network
View
PDF
for article titled, On the Effect of the Activation Function on the Distribution of Hidden Nodes in a Deep Network
We analyze the joint probability distribution on the lengths of the vectors of hidden variables in different layers of a fully connected deep network, when the weights and biases are chosen randomly according to gaussian distributions. We show that if the activation function φ satisfies a minimal set of assumptions, satisfied by all activation functions that we know that are used in practice, then, as the width of the network gets large, the “length process” converges in probability to a length map that is determined as a simple function of the variances of the random weights and biases and the activation function φ . We also show that this convergence may fail for φ that violate our assumptions. We show how to use this analysis to choose the variance of weight initialization, depending on the activation function, so that hidden variables maintain a consistent scale throughout the network.