I am having a hard time understanding the Gradient Descent Rule for learning in a feedforward ANN. In particular, how do we determine the initial weight vector, and how is this weight vector adjusted after each epoch?
From what I've read, I know that we first define some error function depending on the weights, and I think that we choose the initial weight to be the minimizer of this error function. Is this right?