Submitted by These-Assignment-936 t3_10y2mu0 in MachineLearning
I just finished reading the paper "Pre-Trained Language Models for Interactive Decision Making" (https://arxiv.org/abs/2202.01771). As I understand it, the authors are using a language model to "generate" an optimal path to an objective, in test environments like VirtualHome and BabyAI. Reinforcement and imitation learning are evaluated as ways for the model to self-improve.
This is the first time I've seen a language model being used to "solve a problem" that isn't a language one. It seems to open up so many new possibilties. Has this been done before? Are there other examples of LMs being used as decision engines? What's the state of the art? Any interesting applications you've seen?
Side question: I imagine there were AI approaches to navigating VirtualHome and BabyAI that were NOT language-model based. What is the standard modeling approach to these kinds of problems?
currentscurrents t1_j7wf3u0 wrote
>What is the standard modeling approach to these kinds of problems?
The standard approach is reinforcement learning. It works, but it's not very sample-efficient and takes many iterations to train.
LLMs are probably so good at this because of their strong meta-learning abilities; during the process of pretraining they not only learn the task but also learn good strategies for learning new tasks.
This has some really interesting implications. Pretraining seems to drastically improve sample efficiency even if the pretraining was on a very different task. Maybe we could pretrain on a very large amount of synthetic, generated data before doing our real training on our finitely-sized real datasets.