Imagine we have pictures of three animals: dogs, cats, and horses. We train our image classifier and get a confusion matrix, noticing that the model tends to predict that dogs are horses.
But then we read Cross Validated and learn that threshold-based scoring rules like accuracy have serious flaws, so we want to look at the predicted probabilities, not just take the category with the highest probability.
Is there a way to adopt the confusion matrix to the probability outputs to notice that the model tends to believe that dogs are horses? The idea that comes to mind is to take the sum of the predicted probabilities (which works out to be the same as a confusion matrix, if we do this to predicted probabilities that we "round" to give a probability of $1$ of the category with the highest probability). Has this been explored in any literature?