This may be a silly question, but I can't find a concise answer. I've been studying Convergence of Random Variables in Wasserman's All of Statistics, which starts out by explaining:
$X_n$ is a sequence of random variables: $\{X_1, X_2, ... X_n\}.$
It then goes on to define the law of large numbers:
$\bar{X_n} = n^{-1} \sum_{i=1}^{n} X_i$ converges in probability to $\mu = \mathbb{E}(X_i)$
This seems like a mean to me, however I'm struggling to understand this intuitively. Am I supposed to be visualizing the addition of random number distributions, which could eventually end as a point-mass at the expected value? But couldn't $\mathbb{E}(X_i)$ resolve to many possible numbers since the relatedness of the sequences of $X$s is unknown?
+
=
?
Or am I taking just a sample from each distribution to add?
$\bar{X_n} = n^{-1} \sum_{i=1}^{n} \{x$ $\epsilon$ $X_i\}$?
My problem is that the concept of adding random variables (as opposed to samples from them) was never defined. I'm assuming I just add the closed form representation of their distributions, but you know what they say about assumptions...
Thanks!