1

I'm not a statistician and this is my first post on stackexchange, so I apologize if I'm not following guidelines.

My situation is that I have a prognosticator who projects sales revenue for 30 different products in my company and I have 3 years worth of data. My goal is to determine what his predictions would look like over 1000 years.

First, I normalized the difference between his predictions and actual results to percentages of the predicted guess, since different products have different projections.

So I'm assuming that the difference between his predictions and actual results look like a normal distribution.

Right now, I'm getting a mean of 2% and standard deviation of 13%.

My question is that I want to see what the normal distribution looks like around a mean of 0. I know that right now, that on average he tends to over predict by 2%, but I'm assuming that due to the small sample size that he could just as like be under predicting by 2%. So I'd like to take the same data and have a mean of 0 and know how I'm supposed to change the standard deviation.

For example, if I shift the mean to 0, would the standard deviation stay 13%? I assume the standard deviation should move up to account for the fact that if we're moving the mean to 0 that we're in essence saying that his predictions are more accurate and therefore the standard deviation would need to be increased to take this into consideration.

Does that make sense? Thx in advance.

Santaur
  • 11
  • 3
  • 8
    You can change the mean to $0$ simply by subtracting the sample mean from all the sample values to create a new set of "centered sample" (meaning that the mean of the centered samples is $0$. The centered samples enjoy the same standard deviation as the original samples. – Dilip Sarwate Aug 23 '17 at 19:12

1 Answers1

1

The standard deviation is calculated by using the empirical mean as the best estimate for the true mean. If you for some reason know the true mean you can calculate a better estimate of the standard deviation as $$\sigma=\sqrt{(1/N)\sum (X-\mu)^2}$$ This can give a higher standard deviation as you mention. But it can also be lower since the unbiased (default) estimate of the standard deviation is $$\sigma=\sqrt{[1/(N-1)]\sum (X-\mu_{emp})^2}$$ You notice here the $1/(N-1)$ is different from the $1/N$: this is to take account of the downward bias added when using a estimated mean value. The empirical mean is the value that minimizes the sum, so by using the empirlcal mean without the correction the standard error measurement will be biased.

Nick Cox
  • 48,377
  • 8
  • 110
  • 156
  • 2
    Using $N-1$ rather than $N$ gives an unbiased estimator of the variance, but the square root of that isn't an unbiased estimator of the SD. The remaining bias is often ignored, but it exists. – Nick Cox Mar 02 '18 at 07:03