Submitted by AmalgamDragon t3_yf73ll in MachineLearning
My search foo has failed me, so I'm wondering if anyone has heard of using a pair of NNs with identical architectures, but different seeds, to train a model for producing embedding vectors without defining a supervised learning task. The outputs of the NNs would be the embedding vector directly. The two NNs would be trained synchronously and use the output of the other as the label with the training continuing until the outputs of the NNs become sufficiently close.
Has anyone heard of such a thing?
DaLameLama t1_iu2jo4l wrote
You need a way to prevent the training from collapsing to trivial solutions (like, both NNs output the same constant vector for all inputs).
Methods similar to your idea are Barlow Twins or VICReg.