I am working on an imbalanced binary classification and wondering in what step I should find the best optimal threshold cut-off point. When I tried classifying the dataset with the normal probability threshold, I couldn't see any true positives(failures) but when I lowered the probability threshold, the model started showing me some true positives.
What I am really confused about finding the optimal threshold cut-off is in what step I should find the optimal threshold cut-off point. The following questions I have in mind:
Do I need to find the optimal threshold cut-off point before the model development?
Do I use cross-validation for it?
Do I need separated procedures for finding the optimal threshold cut-off point from model developments? So I run two procedures, one for the optimal probability threshold and the other for the model development.
I hope to hear some advice or answers.