Consider the posterior distribution $p(\theta|x)$. We aim to find a "good" estimate of the random variable $\theta$. The Bayes risk associated with the loss function $L(\hat{\theta}, \theta)$ is denoted $E(L(\hat{\theta} , \theta)|x)$.
For the mean square error loss function $L(\hat{\theta} , \theta) = (\hat{\theta} - \theta)^2$, a well known result is that $\hat{\theta}=E(\theta|x)$ minimises the Bayes risk (minimum mean square estimator).
But what about the multidimensional case, when we aim to estimate a random vector $\theta=(\theta_1 ... \theta_n)^T$? With $L(\hat{\theta} , \theta) = \sum_i(\hat{\theta} - \theta)^2$, is the best estimate $\hat{\theta}_i = E(\theta_i|x)$?
If it is indeed the case, this is counter-intuitive to me because this estimate would be characterised solely by the posterior marginals, whereas the Maximum A Posterior (MAP) estimator depends on the joint posterior marginal.
Finally: same question for the cost function $L(\hat{\theta} , \theta) = |\hat{\theta} - \theta|$. In one dimension, the minimum is the median of $p(\theta|x)$. In multiple dimensions, with $L(\hat{\theta} , \theta) = \sum_i|\hat{\theta_i} - \theta_i|$ is the minimum obtained at the medians of the posterior marginals?