Viewing a single comment thread. View all comments

Oceanboi OP t1_ixkfcoh wrote

So all we really know is that if a model has been trained on some previous task, there’s some arbitrary probability that it can be used for another problem, regardless of image contents or problem domain?

2

Think_Olive_1000 t1_ixmm3a2 wrote

Yes, but how well it works will be limited by whether you can find exploit a similarity between the tasks.

Tangentially related: when openai were training their speech recognition model 'whisper' they found that when they trained the model to perform translation it also inexplicably increased the models performance in plain english transcription.

1