Suppose I have some reasonably sized dataset and I do linear regression on it, so now I have a model, say $\hat{y}=Ax+b$, where $y$ is real (or perhaps a vector, but let's say real for now), $x$ (my data points) are vectors, and $A$ and $b$ are the parameters of my linear fit that I've found through whatever method.
Now I have a new datapoint (say $x_0$) and I run it through my model, which has a true "output" value of $y_0$ but a predicted value of $\hat{y}_0$. The question is: how precise do I expect this prediction to be? Better yet, can I construct a $95\%$ confidence interval (or whatever percentage) for $y_0$?
I'd be happy to have a link to any sources on this topic; I assume it's complicated enough that it's not a one-line answer. It hasn't come up in the courses I've taken so far, so I wonder if it's even been done (it seems like it would have a lot of requirements, but one might hope the central limit theorem could help reduce the number of required hypotheses).
EDIT: I have, through the magic of tags, learned the word "prediction interval" and suspect it may be what I want. But I do not know anything about them, or how to compute them, or what we assume about the data to make them meaningful.