Suppose there is a set of $n$ independent observations $y_i$ from the exponential family of distributions. How can we prove that in a saturated GLM model, the fitted MLE values $\hat\mu_i = y_i$ for all $1 \leq i \leq n$?
Edit: Using the new definition proposed I am still not sure I get the proof correct. Let $\dot f$ denotes the derivative of $f$ with respect to the canonical parameter $\theta$, $g$ the link function and $V$ the variance function, we know that if $D = \mathrm{diag}\left( [\dot g(\mu_i)V(\mu_i) ]^{-1} \right)$ is a $p\times p$-matrix then it is easy to see that $D$ must have full rank since it is diagonal with non-zero columns. Maximizing the likelihood means $X^TD(y-\mu) = 0$. If this arises from a saturated model, then by the answer below, $\rm{rank}\ X = p$, We know that the product of two matrices of full rank has full rank, and therefore if we can somehow see that $X$ is square, the invertible matrix theorem tells us there exist an inverse, C, for $X^T$ such that $$ D^{-1}CX^TD(y - \mu) = D^{-1}C 0 = 0 \qquad \text{only if} \qquad y - \mu = 0. $$ So is there something that forbids linearly dependent columns of $X$? Can $X$ be singular? As far as I can tell, it would nail the proof. What do you think?