Submitted by windoze t3_ylixp5 in MachineLearning
Hey, I'm a casual observer of the DL space, what are the biggest technique changes or discoveries that are now used everywhere? From my view:
- Pretraining - reuse large data sets in the same domain (2010)
- ReLU - simple to train non-linear function (2010)
- Data Augmentation - how to make up more data (including noise, random erasing) (2012-)
- Dropout - how to not overfit (2014)
- Attention - how to model long range dependencies (2014)
- Batch normalisation - how to avoid class of training issues (2015)
- Residual connections - how to go deeper (2015)
- Layer normalisation - how to avoid class of training issues (2016)
- Transformers - how to do sequence modelling (2017)
- Large Language Models - how to use implicit knowledge in language (2019)
What's the other improvements or discoveries? More general the idea the better.
Edit: added attention, pretraining, data augmentation, batch normalisation, contrastive methods
ukshin-coldi t1_iv0593t wrote
Your dates are wrong, these were all discovered by Schmidhuber in the 90s.