I have a student (for a high school project) who has conducted an experiment and collected data in the following (simplified) way:
id|situation|scoreX|scoreY
where id
identifies the subject; situation
only takes the nominal values A
or B
; and scoreX
and scoreY
are numerical measurements. Each subject has two entries, one where situation=A
and the other when situation=B
.
Originally she planned two hypothesis tests, using a t-test, with the alternatives: mu_scoreX_A > mu_scoreX_B
and seperately mu_scoreY_A < mu_scoreY_B
.
My thinking is that if the two original tests were to give p-values of 0.1, individually they are not significant, yet we are seeing two 'almost' effects, which together could be significant.
So my question: isn't it better to do one test against the alternative: mu_scoreX_A > mu_scoreX_B AND mu_scoreY_A < mu_scoreY_B
. If so, what is the test one should use and do we need to adapt the significance too?
Taking my thinking one step further, are there rules and tests for any boolean combination of alternatives?