In my analysis I am testing a lot of similar models. I have two variables (let´s call them A and B) and a whole bunch of other variables (C1....C10). All models have the same approach: It is always looking whether there is an interaction effect Between A or B with one of the C1..C10. So for example:
- model 1: interaction A with C1
- model 2: interaction A with C2
- ...
- model 10: interaction A with C10
- model 11: interaction B with C1
- model 12: interaction B with C2
- ...
- model 20: interaction B with C10
I can´t combine all the C variables because they are sometimes quite similar and therefore I would expect them to cause multicollinearity. Combining A and B would lead to a lot of NAs. For those two reasons I separated the models. On the other hand I now have 20 different results and I am not sure how to interpret them correctly. My approach would be to look at the p-values (as it is an easy criterion) and assert that those with a low p-value probably show some correlation (have a significant interaction effect) whereas models with high p-values appear not to have significant interaction effects and show contrary results to what I expected.
- Is this approach OK?
- I do not omit any results. I will show those results with high p-values as well. As far as I understood "p-hacking", it is similar to "cherry-picking". I therefore should not encounter such problems or am I wrong?