I am training an ML model for a supervised classification problem. For this, I have been provided with two datasets. They both contain the same number of samples, however dataset 1 includes data for 14 features whilst dataset 2 includes data for 46 different features.
The question I want to answer is which dataset can be used to train a better model. However, I am wary that the datasets are "imbalanced" with regards to numbers of features, and am not sure whether I need to correct for this in order to make a fair comparison between the two. Specifically, the hypothesis we have is that the 46 features in dataset 2 are individually more informative than the 14 features in dataset 1, and so a "feature balanced" dataset 2, formed from 14 features of the original dataset 2, would be able to be used to train a better model than dataset 1. However, I don't want to throw away data unnecessarily. Therefore, I was wondering if I need to perform such "feature balancing" or not in order to fairly make this comparison.