I am a programmer with little mathematical background who started to study statistics/ML recently. I quickly stumbled upon the random variable term and it was hard for me to understand why in statistics it is called random. For me a truly random variable must be impossible to predict, in other words it must have uniform distribution. If a variable is not uniformly distributed we can somewhat predict it. In fact in programming a random number generator is considered bad if distribution of it's generated values is not quite uniform. So I was surprised to see that we can build a good model and predict "random variables" in statistics. Looks like terminology clash to me.
I just want to check if my understanding of the random variable is correct. So my questions are:
- Is it true in statistics that any variable which values cannot be precisely predicted is called random variable regardless of it's distribution?
- If we have a good (but not perfect) model for predicting a random variable $Y$ is it still random? Or let me try to say it more formally. if we have some dependence $Y = f(X) + \epsilon$ and a model $\hat{y} = \hat{f}(x)$ which has little but non-zero error on both test and training samples is $Y$ still a random variable? Here the $\hat{y}$ is a predicted (but not observed) value of the variable $Y$; $\hat{f}$ is an approximation of $f$; $x$ is any value of independent variable $X$ and $\epsilon$ is an irreducible error variable which depends on some unknown events (basically the deviation of $Y$ around the $f(X)$).
Thank you.