Why do we use SSE (the sum of squared errors) in calculations. Why do not we use the sum of errors? Could you explain please?
Asked
Active
Viewed 2,167 times
0
-
1I presume you meant "sum of absolute errors" for `sum of errors`. – ttnphns Apr 19 '14 at 14:09
-
1The main reasons are two: 1) SSE is related to the estimating of the _mean_ (See closely [related question](http://stats.stackexchange.com/q/81986/3277)); 2) SSE and the mean are naturally related to _linear algebra_ computations (see e.g. [this](http://stats.stackexchange.com/a/14211/3277)). Still, absolute error is used sometimes too (for example when the central tendency is median). – ttnphns Apr 19 '14 at 14:15
-
1Basically a duplicate of [this](http://stats.stackexchange.com/questions/48627/sum-of-squares-of-residuals-instead-of-sum-of-residuals?lq=1), where in turn [this](http://stats.stackexchange.com/questions/46019/why-squared-residuals-instead-of-absolute-residuals-in-ols-estimation) is linked as 'essentially a duplicate'. Another near duplicate is [this](http://stats.stackexchange.com/questions/118/why-square-the-difference-instead-of-taking-the-absolute-value-in-standard-devia?lq=1). There are others as well. – Glen_b Apr 19 '14 at 17:14
1 Answers
2
Errors can be both positive or negative. So if you just add them up (without squaring them), then the positive terms will cancel out the negative terms. This means that you are actually ignoring the errors that you have in your fitted model which is not good. But if you square them and find sum of squared errors, all the errors (i.e. both negative and positive) are taken into account since squaring makes all of them positive.

Stat
- 7,078
- 1
- 24
- 49