Which of the two methods is the most effective to avoid overfitting: regularization or searching the model's hyperparameters through crossvalidation? is one of these methods preferable for large/small sets of data? is there any experimental evidence concerning this question? Thanks. |
The 2 methods are generally used together since they have different purposes:
It is possible to combine the 2 together by doing a cross validated grid search for the optimal value of the regularization parameter (e.g. C in SVM). Edit: arguably if the algorithm is able to scale to large datasets (linear training time) and that this data is available cheaply (generally not true for supervised learning) then regularization is less important as the redundancy in the training set will generally act as a natural regularizer that will prevent over-fitting. It is still interesting to do cross validation (maybe online cross validation to make it scalable) so as to measure the remaining amount of overfitting. |