1

I guess I could find the answer to my question if I knew the right Google search words.

If you use a better model in a classification problem then you will get a better accuracy (if you use that metric to judge). But even the "best" model will reach a ceiling, e.g. if the data is not telling you enough about the problem.

Is there a way to somehow separate the "essential accuracy" (the max accuracy the best model could achieve) from the "incidental accuracy" (the accuracy you happen to achieve due to your potentially imperfect model)?

Which search terms would I have to use to learn more about this topic?

Thanks! Christian

cs224
  • 111
  • 2

1 Answers1

0

After reading more around this topic I believe that the paper Understanding predictive information criteria for Bayesian models answers my original question best and gives great references to further reading. It also clarifies terminology and vocabulary and gives some concrete examples and comparisons.

In the end, my conclusion is that my question was really around having a "measure of predictive accuracy" and comparing that to the "true data-generation processes" "measure of predictive accuracy". The paper shows what is possible and what is not.

cs224
  • 111
  • 2