Submitted by Lajamerr_Mittesdine t3_ycipui in MachineLearning
say_wot_again t1_itrmhsx wrote
Reply to comment by DeezNUTSampler in [R] Large Language Models Can Self-Improve by Lajamerr_Mittesdine
Here's an example of what I had in mind. Pseudolabels for unlabeled data are generated on the clean images, but the student model is trained on a strongly augmented version of the image. It's not contrastive learning because the objective is still explicitly object detection, but instead easy vs hard is the original image vs the strongly augmented one.
Viewing a single comment thread. View all comments