Say I have a random vector $Y\sim N(X\beta,\Sigma)$ and $\Sigma\neq\sigma^2 I$. That is, the elements of $Y$ (given $X\beta$) are correlated.
The natural estimator of $\beta$ is $(X'\Sigma^{-1}X)^{-1}X'\Sigma^{-1}Y$, and $\text{var}(\hat{\beta})=(X'\Sigma^{-1}X)^{-1}$
In a design context, the experimenter can fiddle with the design which will result in different $X$ and $\Sigma$ thus different $\text{var}(\hat{\beta})$. To choose an optimal design, I see that people often try to minimizes the determinant of $(X'\Sigma^{-1} X)^{-1}$, what is the intuition behind this?
Why not, say, minimizes the sum of its elements?