I have a statistical model with around 20 predictor variables, built on 90% of a dataset consisting of over 600k observations. The original developer held out 10% of the original dataset for the purpose of external validation.
From my reading, it seems that even cross-validation is sensitive to the partition of the dataset, let alone this one-fold data-splitting. I was trying to get a more objective measure of the predictive ability of the model that has been developed by bootstrapping.
If I am not concerned about the robustness about the model development procedure, but just interested in quantifying the (estimated) predictive ability of this particular model that was developed (without re-estimating the parameters),
- Would it be appropriate to sample with replacement 10000 observations and compute the c-statistics, repeating the process say 100-500 times?
- What are some potential drawbacks / danger of this (not obtaining sample the same size as the original dataset - as it would be too costly and time-consuming to do)?
- What alternative approach would you suggest?
Thank you very much!
Related stats.stackexchange.com entries: