A reviewer recently took to my paper with a baseball bat suggesting that the 'Table 2 Fallacy' rendered my results essentially uninterpretable.
In the paper I used penalised regression (lasso) to assess what factors were important predictors of meeting criteria for cannabis use disorder. We had seventeen predictors and ~900 observations. The study was cross-sectional.
This was the first I had heard of the Table 2 Fallacy, so I read about it online here and at CV here. It seems that the gist of the argument is that (a) you need some sort of structural equation modelling approach to properly make sense of the relationships between predictors and outcome variables (b) penalised regression and statistical learning techniques are only of use for prediction and not so much for explanation. This was a bit dispiriting to me because I understood that penalised regression was a more ethical way to identify predictors of a particular outcome in the absence of a unifying explanatory model. Also it seems to me that structural equation modelling is only useful with a manageable number of covariates, certainly a lot less than 17. Yet the 17 we identified and added to out model have all been found to be associated with cannabis use disorder throughout the literature.
So my questions are:
1. Is the reviewer right? Does the absence of structural equation modelling render my results meaningless?
2. Under what circumstances would penalised regression with a large number of predictors have validity as a tool for explanation (i.e. rather than merely prediction)? (i.e. how do the myriad of papers using penalised regression get around this problem? Are their better analyses we could conduct?)