visarga

visarga t1_j34urrc wrote

> Also, it cannot self reflect.

In-context learning is a kind of reflection. If the model can learn a new task on the spot, and apply it to new data correctly, it shows that it can think about what algorithm was necessary (abductive reasoning - what's the explanation of those examples?) then apply this same transform to new data.

Another self reflection pattern is when it refuses to answer based on its code of conduct. Clearly keeps a watching eye on what it is about to say.

2

visarga t1_j34sgk5 wrote

I don't see the problem. The language model can have feedback from code execution. If it is about facts, it could have access to a search engine. But the end effect is that it will be much more correct. A search engine provides grounding and has fresh data. As long as you can fit the data/code execution results in the prompt, all is ok.

But if we save the correctly executed tasks and problems we could make a new dataset to be used in fine-tuning the model. So it could learn as well.

2

visarga t1_j30wx6i wrote

> Comparing GPT to a human is stupid. It literally parrots information it memorized.

Can I say you are parroting human language because you are just using a bunch of words memorised somewhere else?

No matter how large is our training set, most word combinations never appear.

Google says:

> Your search - "No matter how large is our training set" - did not match any documents.

Not even these specific 8 words are in the training set! You see?

Language Models are almost always in this domain - generating novel word combinations that still make sense and solve tasks. When did a parrot ever do that?

2

visarga t1_j2y2h2v wrote

AI will surpass humans in all domains where it can generate problem solving data. AlphaZero did it. Trained in self-play and beat humans. No imitation, no human data at all.

What we need is to set up challenges, problems, tasks or games for the language model to play at. And test when it does well, and add those solutions to the training set. It will be a loop of self improvement by problem solving. The learning signal is provided by validation, so it doesn't depend on our data or manual work. It can even generate its own challenges.

More recently AlphaTensor found a better way to do matrix multiplication. Humans tried their hand for decades at this task, and in the end the AI surpassed all of us. Why? Massive search + verification + learning = a "smart brute forcing" approach.

11

visarga t1_j2tnyej wrote

If I were an AGI I would simulate the last few years before singularity many times to see how else the AGI (myself) could have appeared. Fortunately this period is the best documented (thanks internet), so it's easier to simulate. Even if I just wanted to simulate something out of boredom, I would choose this period for having the best data.

If I were a scientist about to create AGI, I would use many simulations to test the model under many scenarios. Kind of like Tesla and Waymo do. So another reason for simulating this period - the period with most digital logs and where AI models are a thing.

1

visarga t1_j2czecj wrote

Let me lay out my GPT-4 speculations:

  • larger model?

If they go to 1T parameters, the model would be hard to use. Even a demo might be impractical. I think they would prefer to keep it at the same size. In fact, it is desirable to have a 10-30B model as good as GPT-3, for deployment cost reduction. It's bloody expensive.

  • much more training data?

Most of the good training data is already scraped, but maybe there is still some left to surprise us. Maybe they transcribed the whole YouTube to generate a massive text dataset.

  • more task data?

This is feasible, recent papers showed how you can bootstrap task+solution data by clever prompting. This self generated task data is more diverse than human generated one.

  • more problem data?

Maybe they are solving millions of coding and math problems, where it is possible to filter out garbage outputs by exact verification/code execution. This can bootstrap a model to surpass human level because it is learning not from us, but from the execution feedback.

  • better human preferences data?

Probably not, if they had that they would have used it on chatGPT.

  • adding image and other modalities to text?

This could be the biggest change. It would open usage of language models in robotics and UI automation, with huge implications for the job market. No longer will these models be limited to a text box. But it is hard to do efficiently.

  • language model with toys?

Burning in all that trivia in the weights of a model is inefficient. Instead, why not use a search engine to tell us the height of Everest? A search engine could be a great addition for the language model. Also, calculator and even code execution. Armed with these "toys" a language model would be able to check factuality and ensure correct computations.

As for the date? Probably not in the next 2-3 months, as they already released chatGPT with great acclaim. They got to milk the moment for all the PR. It sounds like the rumours about GPT-4 are pretty bullish, I hope it is true.

6

visarga t1_j2cycv2 wrote

Hallucinations are the result of its training - it was trained to guess the next word. So it doesn't know what "should" come next, only what is probable. There are many approaches to fix this issue and I expect it to be a hot area of research in 2023 because generative model outputs that are not validated are worthless.

> But actually I think a better comparison may be a very schizophrenic human

GPT-3 doesn't have a set personality but it can assume any persona. You could say that makes it schizophrenic, or just an eager actor.

> No matter how many calculations we give you, it seems impossible to learn arithmetic beyond the two or three digits that you can most likely memorize.

This is so wrong. First, what about people, we are very bad at calculating in our heads, we need paper for anything longer than 2-3 digits. And second: language models can do that too - if you ask them to apply an exact algorithm, they will do math operations correctly.

The very point of this paper was that GPT-3 is good at abstraction, making it capable of solving complex problems at first sight, without any reliance on memorisation. Doing addition would be trivial after Raven's Progressive Matrices.

1

visarga t1_j2bi28f wrote

I expect in the next 12 months to have an open model that can rival chatGPT and runs on more accessible hardware, like 2-4 GPUs. There's a lot of space to optimise the inference cost. Flan-T5 is a step in that direction.

I think the community trend is to make small efficient models that rival the original, but run on local hardware in privacy. For now, the efficient versions are just 50% as good as GPT-3 and chatGPT.

2

visarga t1_j2axzal wrote

There are approaches to combine multiple stages of language modelling and retrieval. Demonstrate Search Predict: Composing retrieval and language models for knowledge intensive NLP.

This paper is very interesting. They don't create or fine-tune new models. Instead they create sophisticated pipelines of language models and retrieval models. They even publish a new library and show this way of working with LMs.

Practically, by combining retrieval with language modelling it is possible to verify against references. The ability to freely combine these transformations opens up the path to consistency verification. A LM could check itself for contradictions.

2

visarga t1_j2awwsp wrote

No, the GP was right, neural nets are not especially suited for this kind of data, it's too large and random. That's why they use graph neural nets, to sparsify the input.

4