Quasilikelihood (like pseudolikelihood, M-estimation, unbiased estimating functions, and composite likelihood) came from the realisation that many of the desirable properties of maximum likelihood estimators could be obtained without using the actual likelihood.
If you have an objective function $Q(\theta)$ that has its maximum at the right place (so $\theta_0$ maximises $E_{\theta_0}[Q(\theta)]$), plus some conditions on smoothness and variance, $\hat\theta$ will be consistent and asymptotically Normal by essentially the same arguments as for a maximum likelihood estimator. So, it's like an MLE except that you don't need to use the likelihood.
There isn't any requirement that the objective function you use isn't a likelihood -- it's just that when it is a likelihood you might not call it quasilikelihood, you might use some other name. Nowadays, I think the distinction between quasilikelihoods and other objective functions that aren't the likelihood is becoming less important
So why not just use the likelihood? Well, you might not know it. If you have, say, count data that are not a good fit to a Poisson distribution, you might want to have a straightforward analysis that doesn't require you finding a distribution that does have a good fit. A quasilikeliood estimator that captures the dependence of the mean $\mu$ on $X$ and the dependence of the variance on $\mu$ will give reasonable estimates and not require inventing a new parametric distribution and then programming it (this was the 1970s, remember)
Quasilikelihood estimators will typically not be fully efficient -- if you did know the likelihood you could do better -- but they may be simple enough or well-behaved enough that you are prepared to live with the efficiency loss. They also don't necessarily satisfy the information equality, so log quasilikelihood ratios don't necessarily have an asymptotic $\chi^2$ distribution (though they do have a known asymptotic distribution)