Disclaimer: I barely know what I’m doing, fact check me.
What you’re describing is called online learning. It can be done, but I imagine it doesn’t work well unless you have a clear signal on what “correct” was supposed to be for an example output.
You could use user feedback as a quality signal, but you need a way to trust that the user feedback is correct enough to integrate as new training data.
This is all probably possible, but it’s layers of added complexity that they might not be interested in right now
asterfield t1_irjc3wd wrote
Reply to [D] Why can't language models, like GPT-3, continuously learn once trained? by SejaGentil
Disclaimer: I barely know what I’m doing, fact check me.
What you’re describing is called online learning. It can be done, but I imagine it doesn’t work well unless you have a clear signal on what “correct” was supposed to be for an example output.
You could use user feedback as a quality signal, but you need a way to trust that the user feedback is correct enough to integrate as new training data.
This is all probably possible, but it’s layers of added complexity that they might not be interested in right now