Questions tagged [weight-initialization]

4 questions
4
votes
0 answers

Are Batch Normalization and Kaiming Initialization addressing the same issue (Internal Covariate Shift)?

In the original Batch Norm paper (Ioffe and Szegedy 2015), the autors define Internal Covariate Shift as the "the change in the distributions of internal nodes of a deep network, in the course of training". They then present Batch Norm as a solution…
1
vote
0 answers

Guide to self-starter estimators (parameter initialization) for "simple" functions

Background I have a collection of functions with trainable parameters that I am implementing as Keras model classes, which enables immediate use of a variety of objective functions, optimizers, and training-related methods (e.g. early stopping…
1
vote
1 answer

Output of ANN with zero initialized weights represents what?

In class we discussed that if the weights of an ANN (standard feed forward NN in binary classification setting [0,1]) are initialized all at zero, the ANN fails to break symmetrie and therefore, the units in each layer develop equivalently. My…
0
votes
0 answers

Correct weight initialization for pre-activation convolutions & pre-activation depthwise separable convolutions?

My CNN architecture use pre-initialization, i.e. BatchNorm -> ReLU -> Conv. Which weight initialization shall I use for the convolutions? I'm under the impression that the standard ReLU initialization scheme of HeNormal is designed for Conv -> ReLU…