I have a simple regression model (y = param1*x1 + param2*x2). When I fit the model to my data, I find two good solutions:
Solution A, params=(2,7), is best on the training set with RMSE=2.5
BUT! Solution B params=(24,20) wins big on the validation set, when I do cross validation.
solution A is sorrounded by bad solutions. So when I use solution A, the model is more sensitive to data variations.
solution B is sorrounded by OK solutions, so it's less sensitive to changes in the data.
Is this a brand new theory I've just invented, that solutions with good neighbours are less overfitting? :))
Are there generic optimisation methods that would help me favour solutions B, to solution A?
HELP!