I'd vote for singular values/eigenvalues/eigenvectors over determinants and adjugates for the way to approach this.
TLDR: standard errors increase as the eigenvalues of $X^TX$ get increasingly small and this corresponds to the formation of valleys in the loss surface representing our increasing inability to separate out candidate $\hat\beta$ values.
We're looking to minimize $\|y - Xb\|^2$ over $b\in\mathbb R^p$. Let $X = UDV^T$ be the SVD of $X$. As $X$ gets increasingly close to reduced rank we'll have $d_p\to 0$ (at least) where $d_p$ is the smallest singular value. This reflects the fact that $X$ is getting closer and closer to having a non-trivial null space, which would include (at least) $\text{span}(v_p)$, with $v_p$ being the smallest right singular vector or equivalently smallest eigenvector of $X^TX$.
This means that once we've got $\hat\beta$ we could get an almost identical loss by replacing $\hat\beta$ with $\hat\beta + \alpha v_p$ for $\alpha \in \mathbb R$. This shows that there is a whole affine subspace of almost equal loss (at least for modest values of $\alpha$), and as $d_p\to 0$ the loss will become increasingly equivalent over that subspace until we are truly unable to pick an element from it since they all have identical loss.
This is one way to picture high variance: when there are very different values of $b$ leading to an almost identical loss, slight perturbations in the data can lead to very different $\hat\beta$s which is basically what high variance means.
This analysis also tells us that, while some individual coordinates of $\hat\beta$ may get high variance, it's really about the coordinates of $\hat\beta$ expressed w.r.t. the basis given by $V$.
Here's an example. I'll build $X$ by picking $U$, $D$, and $V$.
Let
$$
V = \begin{bmatrix}
1 & 0 & 0 \\
0 & 1/\sqrt 2 & 1/\sqrt 2 \\
0 & 1/\sqrt 2 & -1/\sqrt 2
\end{bmatrix}
$$
$$
D = \text{diag}(2, 1.7, .01)
$$
and let $U$ be any matrix in $\mathbb R^{n\times 3}$ with orthogonal columns. This leads to
$$
(X^TX)^{-1} = VD^{-2}V^T \approx \begin{bmatrix} 1/4 & 0 & 0 \\ 0 & 5000 & -5000 \\ 0 & -5000 & 5000\end{bmatrix}
$$
so $\hat\beta_1$ will have a very modest variance but $\hat\beta_2$ and $\hat\beta_3$ have huge variances, and this is because $Xv_3 \approx \mathbf 0$ so $\hat\beta$ can be perturbed along $(0,1,-1)^T$ with only a small change in loss.
So it is true that their individual variances get large but I think this is much more fundamental.