If you compare multiple model fits based on some criterion and then try to compare the pairs of models, then you have the same pairwise comparisons problem that Tukey's method solves in the one-way ANOVA. A major wrinkle here is that the the statistics you are comparing are highly correlated, being based on nearly the same models.
This problem has been solved in the case where the model criterion is correct classification probability, in the following paper:
Westfall, P.H., Troendle, J.F. and Pennello, G. (2010). Multiple McNemar Tests. Biometrics 66, 1185–1191.
In particular, the strong correlation structure between the models that are nearly the same is "harvested" to improve the power of the multiple comparisons procedure.
You could extend that method to apply to likelihood-based tests, but it is new research as far as I know. You need to write the fit statistic as a mean of iid observation-level quantities. Then you can use a multivariate bootstrap to construct multiple comparisons between the means. The bootstrap provides a useful and valid way to incorporate the extremely high dependencies between different models. For the general multivariate idea, see
Westfall, P.H. (2011). On Using the Bootstrap for Multiple Comparisons. Journal of Biopharmaceutical Statistics 21, 1187–1205.
Good luck!