I am trying to understand the significance of the intercept in my model. The model was created using
lm.fit <-lm(target ~ .,train)
and the intercept figures were
Estimate Std. Error t value Pr(>|t|)
(Intercept) -1.783e+02 8.587e+01 -2.076 0.03809 *
Then the model was pruned using
sw.fit = step(lm.fit,trace=0,k=log(nrow(train)),direction="both")
and the intercept figures became
Estimate Std. Error t value Pr(>|t|)
(Intercept) 16.196874 0.282901 57.253 < 2e-16 ***
I understand that in writing out the new formula I would have something like
y = 16.20 + the rest
which is easier to think about than the previous
y = -178 + the rest
It seems as if my accuracy has gone up ( the better P value) but the Std.Error has gone from approx 85 to 0.28 and the T value has gone from approx -2 to 57
I understand that the T value is the estimate divided by the standard error. Thus a smaller standard error causes a larger T-value. I read that
the standard error of a statistic is the standard deviation of its sampling distribution or an estimate of that standard deviation
Why does the pruning decrease the standard error of the intercept?