-
I am getting rather odd results, where our validation data is getting better accuracy and lower loss, than training data. And this is consistent across different sizes of hidden layers. This is our model:
I've tried to remove regularization and dropout, which, as expected, ended in overfitting. I've even tried to decrease the learning rate drastically, with similiar results. Has anyone seen similar results? |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment 1 reply
-
This indicates the presence of high bias in your dataset. It is underfitting. The solutions to issue are:-
When training, a percentage of the features are set to zero (50% in your case since you are using Dropout(0.5)). When testing, all features are used (and are scaled appropriately). So the model at test time is more robust - and can lead to higher testing accuracies.
Apart your specific case, I believe that apart from the Dropout the dataset split may sometimes result in this situation. Especially if the dataset split is not random (in case where temporal or spatial patterns exist) the validation set may be fundamentally different, i.e less noise or less variance, from the train and thus easier to to predict leading to higher accuracy on the validation set than on training. Moreover, if the validation set is very small compared to the training then by random the model fits better the validation set than the training. |
Beta Was this translation helpful? Give feedback.
This indicates the presence of high bias in your dataset. It is underfitting. The solutions to issue are:-
Probably the network is struggling to fit the training data. Hence, try a little bit bigger network.
Try a different Deep Neural Network. I mean to say change the architecture a bit.
Train for longer time.
Try using advanced optimization algorithms.
Make sure the is small class imbalance
This happens when you use Dropout, since the behaviour when training and testing are different.
When training, a percentage of the features are set to zero (50% in your case since you are using Dropout(0.5)). When testing, all features are used (and are scaled appropriately). So the mod…