I have to perform 2 tests of the difference of means for the results obtained by 2 learning algorithms.
TEST 1
- I have 2 algorithms
- I take a data set (with N examples)
- I execute both algorithms 10 times
- Each time, each algorithm is trained on a set of 9/10*N examples and tested against a set of N/10 examples (i.e. a 10-fold cross validation is done), and the result is a XX% value of correct classified examples in such a test set.
Here I use the t-test. I know that the t-test makes assumptions of normality. I use it because the data set is quite big ( N>100 ), so I can assume a normal distribution of data (and if data is normally distributed, the means of a single algorithm are normally distributed, and so the differences of the means of the two algorithsm are normally distributed). Or should I consider only N/10 as the sample size from which the means are drawn?
TEST 2
- I have two algorithms
- I take 10 instances of a combinatorial optimization problem (a Travelling Salesman Problem)
- I run each algorithm on the 10 instances
- For each run, a write down the result value (that is at some distance from the optimal value for that instance)
How can I perform a test on the differences of means? Which means can I consider? Which data should I check for normality?