Here's my situation.
I have a multiple linear regression which I've used to come up with a prediction interval to predict a value y for a given (x1,x2,x3,x4,x5,x6). It reads something like lower: 30, upper:48.
I also have the same exact thing to predict a value y* at another given (x1*,x2*,x3*,x4*,x5*,x6*). It reads something like lower:35, upper:51.
I want to answer this question:
What is the probability that the value y* is greater than the value y?
I think it's a basic question, but I'm not sure.
I could likely come up with this probability if I knew the formula for how the prediction interval is calculated in a multi-variable situation.
Here's what I think should be done, but I wanted to run it by you guys first.
Prediction Intervals are based on a t-distribution with (n-6) degrees of freedom (I have a forced 0 y-int). So I believe the margin of error calculated is then some constant multiplied by the corresponding value from the t-distribution (t_.05/2 with n-6 degrees of freedom). The "some constant" would be the standard error of this particular estimate.
I then just do a basic 2 sample t-test using the point estimate prediction as the means and these constants as the standard errors with my n-6 degrees of freedom. Is this accurate?
Is there a better way?
Thanks