visarga

visarga t1_iu7onr9 wrote

> Tell me the first 10 digits of PI, if you succeed you're a bot.

Tested it on GPT-3, got 3.141592653, anyone here could answer that from memory?

Asked for 100 digits, it spits out this:

> 3.1415926535897932384626433832795028841971693993751058209749445923078164062862089986280348253421170679

checks out, can't be a human unless their name is also Pi or Piscine.

6

visarga t1_iu7nryj wrote

Humans fool and lie to themselves all the time, one thing coming to mind is anti-vaxxers protesting vaccines then still going to the hospital when they get sick, or worse, protesting abortion, and then having one in secret.

Similarly, neural nets will learn the training set perfectly but fail on new data, they give you the illusion of learning if you're not careful. That's why in all papers they report the score on a separate block of tests the model has not seen yet. It's a lying, cheating basterd when it comes to learning. This game AI found a clever way to win points without having to do the whole course.

1

visarga t1_iu7nbwy wrote

Q: Why does a boy dying from an allergic reaction to a bee sting make you cry?

GPT-3:

>There are a few reasons why this might make someone cry. First, it is tragic when anyone, especially a child, dies unexpectedly. Second, an allergy to bee stings is relatively rare, so it may be shocking or upsetting to see someone die from something that seems so innocuous. Finally, the death may be a reminder of our own mortality; even if we are not allergic to bee stings, we could still die from something seemingly innocuous.

Does that pass your test? It's one trial, temp 0.7, first answer that popped up.

In my opinion it's better than average human response.

5

visarga t1_itz7mx1 wrote

Internet use grew by 14x between 1997-2007. Mobile phone users grew by 7x between 2000-2010. Smartphones grew by 12x between 2007-2017. In this time we got e-commerce, social networks, online media, taxi and booking apps, educational materials, open source everywhere, the early discoveries in deep neural nets ... Many or most of these were unexpectedly useful and changed society.

We are in a wild west, 2000's bubble period now with AI. I don't think there will be a crash, it's not that, but I think it will take 10 years to see a profoundly transformed world, and 20 years to go beyond our current horizons.

Who will become the rulers of this new era? People like to bet on big corporations because they got the hardware, money and brains. But I think it's misleading. You can run a model on your computer but you can't run 'a Google' on your computer, it will force you to disclose your private data to use it.

But it's possible that AI models will democratise access compared to the centralised internet services. You can install SD or a language model on your own box in privacy. You don't need to wade through spam, you can chat your questions directly to a polite and knowledgeable assistant. Don't need to see any original site at all, or be online for that matter. It's all in the model and maybe a curated corpus of additional content sitting on your drive. Nobody knows what you're doing and they can't put ads in your face. You don't even need to know how to code or know about AI, because its interface is so natural everyone can use it, and use it for new things without needing to reprogram it.

I just described a lifestyle where humans are surrounded by a loyal AI cocoon, a private space for dreaming and creativity that seems to be on the verge of extinction today. That's my dream, what I want to see.

5

visarga t1_itx6vs1 wrote

They use a large context model to learn (distill) from the gameplay generated by other agents. They put more history in the context so the model needs less samples to learn.

This is significant for robots, bots and AI agents. Transformers are found to be very competent at learning to act/play/work relative to other methods, and this paper shows they can learn with less training.

9

visarga t1_ittgjnj wrote

Just imagine 25 years ago, could you have predicted the explosion of work related to the internet? There's development, content creation, communications, commerce and education. Practically it's a double of the physical world. It made us more efficient by a large margin and yet here we are, employed with jobs. Even delivery people and taxi drivers get jobs from the internet.

How is that logic "automating even part of a job leads to layoffs" standing up to the test? I think the correct answer is that we scale up work to match the available capacity instead of firing people. Our desires scale up faster than automation or resources.

1

visarga t1_ittbiag wrote

> Is there going to be enough companies left doing things 'the old way' to keep employment numbers up even though it's less cost effective?

In the medium term there will be new jobs and applications that were impossible before. A company should expand and diversify instead of firing their people, if they care about profits that is. We also have to tackle global warming and other pesky problems on a grand scale. In the long term I think we'll be post scarcity by a combination of automation and smart materials.

5

visarga t1_itr1f9w wrote

It's gonna be both the good and the bad and some surprising bits we never even imagined. But on the whole I think generative AI has given a wide empowerment to people. AI is more accessible than desktop apps and even mobile apps. You can just talk to it, don't even need to read. It helps developers with snippets of code. It helps artists generate stunning images. But it's not hard to learn, it lowers the entry barrier. It basically adds a few IQ points to everyone who uses it. It will be what Google should have been before it choked on spam and ads - a way to make all information more accessible and reusable. It will also run on your own machine, in privacy.

5

visarga t1_itkcw8e wrote

Any of the recent LLMs can blow away the commercial voice assistants we have today. But why are voice assistants so primitive? It's probably too expensive to give everyone GPT-3 powers in their speaker, but that should change fast because they have models 50x smaller with comparable quality.

But probably that was not the only reason LLMs are not in voice assistants, I bet they are afraid it will be prompted into saying racist things and make bad PR. Who wants a 'MS Tay' on their hands?

3

visarga t1_itil03g wrote

You don't program AI with "statements", it's not Asimov's positronic brain. What you do instead is to provide a bunch of problems for the AI to solve. These problems should test the alignment, fuzz out the risks. When you are happy with its calibration you can deploy it.

But an interesting and recent development - GPT-3 can simulate people in virtual polls. Provided with the personality profile, it will assume the personality and answer the poll questions from that perspective.

>GPT-3 has biases that are “fine-grained and demographically correlated, meaning that proper conditioning will cause it to accurately emulate response distributions from a wide variety of human subgroups.”

Apparently GPT-3 not only is aligned with humans in general, but it is precisely aligned with each demographic. So it knows our values really well.

The problem is now we have to specify the desired bias we want from it and that's a political problem, not an AI problem. It is ready to oblige and have the bias we want, it's even more aligned than we want, aligned to our stupid things as well.

3

visarga t1_itij9o0 wrote

People with skills, no job and lots of needs to fulfil have to become self reliant, alone or in a larger group. Fortunately there are going to be great technologies to support that - solar panels, water filtration, automation and of course AI. A community can organise their own school, clinic and store, build their own houses, repair their equipment. The total dependency on the outside should be diminished.

Of course they can't make their own computer chips and are going to use open source software, same for medicine and construction materials. And they need some initial capital. But that should provide a way for people to use their hands to improve their situation. Own your means of production so you don't need a job. We don't need corporations to UBI us, just our share of resources to build with.

3

visarga t1_itgu5bi wrote

Not exponential, let's not exaggerate. It's quadratic. If you have a sequence of N words, then you can have NxN pairwise interactions. This blows up pretty fast, at 512 words -> 262K interactions, at 4000 words -> 16M interactions. See why it can't fit more than 4000 tokens? It's that pesky O( N^2 ) complexity.

There is a benchmark called "Long Rage Arena" where you can check to see the state of the art in solving the "memory problem".

https://paperswithcode.com/sota/long-range-modeling-on-lra

1

visarga t1_itgs3bn wrote

People working on AI projects also don't know how they will turn out, it's alchemy. I mean, who among the AI community predicted Alpha GO, GPT-3 and Dall-E? Nobody. Being an expert in the field did not mean they knew what was around the corner.

1

visarga t1_itgrbjt wrote

Before that there will be AI that can do smaller parts of a movie - background, costumes, music, short pieces of the dialogue, etc. A human can combine them into a coherent movie. Just like Codex is useful to write code but can't write a whole project on its own.

2

visarga t1_itgqoj0 wrote

There are workarounds for long input, one is the linear transformer family (Linformer, Longformer, Big Bird, Performer, etc), the other is the Perceiver, who can reference a long input sequence using a fixed size transformer.

2

visarga t1_itc1tbb wrote

U-PaLM and Flan PaLM come to mind.

The first one shows we were noising the data incorrectly, different kind of noising has major benefits. Second shows that training on thousands of tasks boosts the model capability to follow instructions and also get better score. So maybe OpenAI had to change their plans midway.

It's also possible that they don't want to scale up even further because it's very impractical. Too expensive to use, not just to train. And recent models like Flan get GPT-3 scores on many (not all) tasks with just 3B parameters.

There's also a question about training data - where can they get 10x or 100x more? I bet they transcribe videos, probably all videos that can be accessed. Another approach is to use raw audio instead of text, works well. I bet they have a large team just for dataset building. BLOOM was managed by an organisation with 1000 people and a lot of their effort was into the dataset sourcing, trying to reduce biases.

33