0

Why is backpropagation not incremental?

The definition of incremental would be to be able to update the weights with every single new data point.

However, in stochastic gradient descent with a batch size of 1, isn't backpropagation incremental? If so why are there implementations of incremental backpropagation?

Firebug
  • 15,262
  • 5
  • 60
  • 127
aceminer
  • 813
  • 1
  • 9
  • 20

1 Answers1

1

I never heard the definition of "incremental". Where do you get the "backpropagation not incremental"? Could you give the reference?

I am pretty sure that, for stochastic gradient descent (SGD), we can update the "weights" from one single data point. Details can be found in following post. The post is not about SGD on neural network, but linear model.

How could stochastic gradient descent save time comparing to standard gradient descent?

There is no fundamental difference between SGD in neural network and linear model. The only difference is that for linear model, gradient has a closed form solution, but backpropagation algorithm is used to calculate gradient in neural network.

Haitao Du
  • 32,885
  • 17
  • 118
  • 213