I'm trying to build a two class classifier on a dataset of around 570 samples. Im evaluating several classificiation stratigies (LDA, QDA,RDA, logistic, logistic with some additional ellements like splines ...) I have difficulties if I should take out a 25% of my 570 samples as a test set or if should just rely on 10 fold crossvalidation to estimate my test error.
I know this subject has been discussed before here on stackexhange but Im still confused when I read te answers.
What I'm thinking right know is that I have to take out 25 %, keep it aside and build my models using the other 75% of the data (so trainingdata). Also perform 10 fold crossvalidation on this trainingdata and use this cross validation error to decide between several classification methods/models.
After selecting a final model, do a last check with the test data (that has not been used otherwise) to do a last performance check of my data. I can not use this testdata to compare classification methods.
Am I correct here?
best regards