is it true that:
After training a multilayer perceptron with one hidden layer using gradient descent, we always get the same decision boundary regardless of the initialization point.
is it true that:
After training a multilayer perceptron with one hidden layer using gradient descent, we always get the same decision boundary regardless of the initialization point.
NO.
With one hidden layer, the objective function is not convex (even for squared loss in regression). Therefore, we may have many local minima instead of one global minima.
Using gradient decent and where to end, will depend on the initialization point.