1 . So, when I get an output of a mixed model, in any statistical package, I get the list of coefficients with its p-values. Are they Wald's?
Yes, generally they are. They may be $Z$-statistics/tests (i.e., assuming that the sample is big enough so the standard errors have no uncertainty) or $t$-statistics (allowing for the uncertainty in std err due to finite sample size); this is usually indicated in the column names (and by the appearance of a "df" or "ddf" [(denominator) degrees of freedom] column in the output).
In your second case (results of "ANOVA"), it's hard to know without reading the documentation exactly what tests are being done. It might be either Wald or LRT and might do some sort of finite-size correction or not (see details under #2).
- When I do ANOVA on such model output, it performs a joint test on all coefficients belonging to a single effects, thus, ANOVA gives me p-value for the main effects, one by one. In certain packages, like R or SAS one can choose among: LRT, Kenward-Roger, [Sattherthwaite], F test, Chi2 test. Which one is the Wald? Is this the Chi2 test? Is LRT the F test? Is the Kenward-Roger a "small-data adjustment" to the Wald?
This is a little complicated.
- Wald tests in general assume the log-likelihood surface is quadratic.
- They may ignore the finiteness of the data set, in particular the uncertainty associated with nuisance parameters such as the residual standard deviation (in which case they are "Wald chi-square tests", because the test statistic is $\chi^2$ (or scaled $\chi^2$) distributed in this case
- If they take the finiteness of the data set into account, they are "F tests" ($F$-distributed test statistic)
- if the experimental design is balanced and nested, the denominator degrees of freedom (df) for the F-statistic can be computed exactly
- if not, then some approximation such as Satterthwaite or Kenward-Roger must be used (so the answer to your question "is the K-R a 'small-data adjustment' to the Wald?" is "yes")
- The likelihood ratio test (also "LRT", or (also) "Chi2", because the test statistic of a LRT is $\chi^2$-distributed: in R, if the output says just "Chi2" and not "Wald Chi2" it's probably a LRT) accounts for the non-quadratic shape of the log-likelihood surface, but not the uncertainty of nuisance parameters due to finite size. Finite-size corrections to the LRT are complicated and rarely used.
- Briefly - which of them (Wald, t-test, F test, Chi2 test, LRT) refer to the model coefficients and which of them refer to the main effects? And which can, if any, refer to both?
I'm guessing that what by "coefficients" you mean tests of single coefficients (e.g. the slope in a regression models) vs. joint tests of multiple coefficients simultaneously equaling zero (e.g. the effect of a categorical predictor with >2 levels). t- and Z-tests specifically refer to single coefficients. F and Chi2 essentially test sums of squares of scaled coefficients, so can refer to single or multiple-coefficient tests. Wald and LRT refer to assumptions about the shape of the log-likelihood surface, so are not specific to single- or multiple-coefficient tests.
See also: GLMM FAQ on denominator df; the "pvalue" help page for lme4; How can I obtain z-values instead of t-values in linear mixed-effect model (lmer vs glmer)?
Corrections and comments welcome.