Submitted by AmalgamDragon t3_yf73ll in MachineLearning
My search foo has failed me, so I'm wondering if anyone has heard of using a pair of NNs with identical architectures, but different seeds, to train a model for producing embedding vectors without defining a supervised learning task. The outputs of the NNs would be the embedding vector directly. The two NNs would be trained synchronously and use the output of the other as the label with the training continuing until the outputs of the NNs become sufficiently close.
Has anyone heard of such a thing?
A_Again t1_iu23iwf wrote
Hello friend, using the contrast in labels without supervised labels usually is referred to as a "contrastive predictive coding". Contrastive because you're using contrast between embeddings, predictive because...well you get it
For your consideration though the term "triplet loss" applies what you're describing...just in a supervised mode. And dual-encoder retrieval systems are used for semantic search
I'm on a bus rn and gotta run I hope these at least help you in your search :)