Submitted by Beneficial_Law_5613 t3_zzqzoy in MachineLearning
Some time ago I saw an article saying it is not preferred to use dropout and any kind of normalization(like batch or layer) in a model. But I am not sure why. Any suggestion about that?
Submitted by Beneficial_Law_5613 t3_zzqzoy in MachineLearning
Some time ago I saw an article saying it is not preferred to use dropout and any kind of normalization(like batch or layer) in a model. But I am not sure why. Any suggestion about that?
Pyrite_Pro t1_j2d64e4 wrote
There is no definitive answer to this question. That’s why the field of machine learning has so much empirical experimenting. I suggest to try whether it improves performance or not.