The idea is to show that if we standardize $Y$ then this function has all properties of a score function needed to develop asymptotic theory. Then, we can just treat the standardized function as the score process and work backward (by integrating) to get the"quasi-likelihood" since the score function is the derivative of the likelihood as you point out.
So with that logic in hand, all we need to do is check the desired properties of a score function. What properties are these?
- We need the expectation to be 0 under $\theta_0$ or in this case the true mean, $\mu$. i.e. we have found the maximize of the likelihood.
- The Fisher information gives us the variance of the score function, $Var(l'(X|\theta_0))=I(\theta_0)$. Where $I(\theta_0)=-\mathbb{E}[l''(X|\theta_0)]$, i.e. the expectation of the second derivative gives us the Fisher information.
So let us prove these properties hold.
Consider the following score standardizion of $Y$,
$$q = \frac{Y-\mu}{\alpha V(\mu)}$$
We will treat this random variable as our score function as we prove these properties.
$$\mathbb{E}[q]=\frac{\mathbb{E}[Y-\mu]}{\alpha V(\mu)}=0$$
$$Var(q) = \frac{\mathbb{E}[(Y-\mu)^2]}{\alpha^2 V(\mu)^2}=\frac{\alpha V(\mu)}{\alpha^2 V(\mu)^2}=\frac{1}{\alpha V(\mu)}$$
$$-\mathbb{E}[\frac{\partial q}{\partial \mu}]=-\mathbb{E}[\frac{\partial}{\partial \mu}V^{-1}(Y-\mu)/\alpha]=-\mathbb{E}[-V^{-1}/\alpha]=\frac{1}{\alpha V(\mu)}$$
So we have established that $q$ may be thought of as a score function. Now let us simply recover the quasi-likelihood for a single observation via integration,
$$Q(\mu|y)=\int_y^\mu \frac{y-t}{\alpha V(t)}dt$$
Finally, by independence, we may write the likelihood for the full data as,
$$Q(\mu|Y)=\sum_{i=1} Q(\mu_i|y_i)$$
So for some parametric distribution, we may apply this by determining what $\mu$ and $V(\mu)$ are and plugging them in, and integrating to get our quasi-likelihood. Finally, I recommend this book for all its uses and generalizations.