Suppose I split my data into two parts -- a training set (having 80% of my data) and a testing (20%) set. I train a model on my training set, and test it on the test set.
What do we learn from predicting on the test set? Are we just looking for a measure of generalization error (and perhaps noticing where it makes mistakes) or is there more information we can get out of it? What can we do with this information?