blazejd OP t1_ix7jpz2 wrote
Reply to comment by Nameless1995 in [D] Why do we train language models with next word prediction instead of some kind of reinforcement learning-like setup? by blazejd
This is interesting, but I was thinking a bit more high-level. In essence, BERT and GPT are both self-supervised language models trained on passive data with a similar objective.
Viewing a single comment thread. View all comments