To solve imbalanced data, I used oversampling strategy using ROSE algorithm in Python. As you may know, ROSE is a smoothed bootstrapping method and we can control the dispersion of the augmented data.
I am wondering — is there any rule of thumb on the dispersion (or shrinkage parameter in Python imbalanced learn (imblearn
) package? We would like to publish a paper on this, so I would probably need to justify the dispersion of augmented data.