Adding to the other answers: You cannot in general show that, $(X^T X)^{-1}$
go to zero when $n \rightarrow \infty$. You would need more assumptions, and you have not specified those. As a simple example, let the model be a one way ANOVA comparing $p$ groups, coded as dummy variables ($p$ dummys without an explicit intercept). Let the number of observations in group $i$ be $n_i$ with $n_1+n_2+\dotsb+n_p$. Then the design matrix $X$ becomes
$$
X=\begin{bmatrix} 1 & 0 & 0 &\dots & 0 \\
1 & 0 & 0 &\dots & 0 \\
\dots \\
1 & 0 & 0 & \dots & 0 \\
0 & 1 & 0 & \dots & 0 \\
\dots \\
0 & 1 & 0 & \dots & 0 \\
\vdots \\
0 & 0 & 0 & \dots & 1 \\
\dots \\
0 & 0 & 0 & \dots & 1
\end{bmatrix}
$$
with $n_1$ rows in first block, and so on. Then $X^T X$ becomes a diagonal matrix with the $n_i$'s along the diagonal, and its inverse diagonal with $1/n_i$ along the diagonal. If now you only can get five observations from the first group, but the other $n_2, n_3, \dotsc, n_p$ all increases to infinity with $n$, then the limit of $(X^T X)^{-1}$ becomes the diagonal matrix
$$
\begin{bmatrix}
1/5 & 0 & 0 &\dots \\
0 & 0 & 0 &\dots \\
0 & 0 & 0& \dots \\
\vdots \\
0 & 0 & \dots & 0
\end{bmatrix}
$$
which is not the zero matrix.
So, in general we can assume the model $y_i = x_i^T \beta + \epsilon_i$ where the disturbances $\epsilon_1, \dotsc,\epsilon_n$ are iid random variables from some distribution with zero mean and common variance $\sigma^2$. In matrix form we can write this model $ Y= X\beta+\epsilon$ and we can ask about the estimate of some contrast of the parameter vector $\beta$, say $c^T \beta$ defined by the contrast vector $c$. In our anova example, the mean of group $i$ is given by the contrast $c^T \beta$ with $c=e_i$, $e_i$ the unit vector with a one in position $i$. So the mean of the first group is the contrast $e_1^T \beta$. I that example the variance of the (least squres) estimate of the contrast $c^T\beta$, $c^T \hat{\beta}$, will go to zero with $n$ for some contrast vectors, and not for others.
So, in general we can ask ways of characterizing those contrast vectors $c$ such that the limiting variance is zero, where the variance of the estimated contrast is
$$
\text{Var}(c^T \hat{\beta})=\sigma^2 c^T (X^T X)^{-1} c
$$
or for conditions guaranteeing that the limiting variance is zero for all contrast vectors $c$ (that will correspond to the original question asked here). One such condition could be that the rows $x_i$ of the design matrix $X$ is obtained as an iid sample from some common distribution (with some necessary conditions on that common distribution, no components can have zero variance, for instance).
There is a paper dedicated to giving such conditions with much detail:
Chien-Fu Wu: "Characterizing the consistent directions of least squares estimates", the annals of statistics, 1980, vol 8 No 4 789--801
http://projecteuclid.org/euclid.aos/1176345071