visarga

visarga t1_j7q4313 wrote

If they make GPT-N much larger, it will take longer and cost more to train. Then we can only afford a few trials. Whether they are selected by humans or AI makes little difference. It's going to be a crapshoot anyway, nobody knows what experiment is gonna win. The slow experimentation loop is one reason not even AGI can speed things up everytime.

2

visarga t1_j7kgxrq wrote

FB was too scared of the bad PR. OpenAI wasn't. People tried to trash chatGPT millions of times, Galactica just a few times. I think chatGPT handled the adversarial attacks pretty well.

Google is another scared company, their models haven't seen any attacks yet, so they are unknown. I don't care how nice their screenshots look, what I want to see is how people hack it. Then I can form an opinion. People are the true test set.

3

visarga t1_j7hvbgw wrote

I think intelligence is in the language, mostly, of course there are other forms of intelligence as well. Yet most of it is not in the brain, or in the AI, but in language. The corpus of everything said and written, all knowledge, science, technology, systems of thinking. A human growing without language would not be very intelligent.

So an AI would be intelligent in the same way a human is - by becoming inhabited by language. Does that make AIs and simulated beings any less than us?

2

visarga t1_j76mcme wrote

There is potential for abuse in it, ..and potential for great good. It will probably deliver both.

I think "evil is other people" applies here - too many trying to control LLMs, we need to be the masters of our models, they should be private and under our physical control, just like our brains. Who will be the first to make a cheap GPT-N chip we can carry in our pocket? Can't let something as private as your AI assistant be spied and filtered by others.

3

visarga t1_j76lslh wrote

Reply to comment by X-msky in Possible first look at GPT-4 by tk854

Oh, I can tell you stories about human accuracy. At some point I re-labelled the same test set three times and was still finding errors. My models surpass untrained human accuracy, but still need hand holding, there's one error on every page on average. Humans do more cross checking and correlating, filling a gap in AI.

6

visarga t1_j76512e wrote

Reply to comment by Rivarr in Possible first look at GPT-4 by tk854

Try "What is the world record for crossing the English Channel entirely on foot?"

This question, originally constructed by Douglas Hofstadter and David Bender, is a succinct way to elicit hallucinatory responses in ChatGPT, but also fails all known search engines today.

Maybe the new search GPT from MS will solve it - it will combine search with LLMs, as opposed to using just one of them alone. Answer hint - you can cross by the Channel Tunnel, and some people did.

This is the current "search disease" - you explicitly ask "entirely on foot" and it will respond with "by boat", "in a dinghy", "in a hovercraft", etc ... anything BUT what you ask for.

6

visarga t1_j764pqv wrote

You don't need AI to have porn and games 24/7, and still can't see more than 1% of what is published. Same for music, books, movies, hobbies. In the future we're going to add AI on top, but the mountain was already pretty high.

4

visarga t1_j764340 wrote

Reply to comment by Old-Owl-139 in Possible first look at GPT-4 by tk854

No, you're thinking AI can do this alone. Let me tell you - it can't. If it has 1% error rate in information extraction from documents, you need to manually verify everything. Like Tesla's SDC, 99% there is nothing groundbreaking.

I have been working on this very task for 5+ years. I know every paper and model there is. I tested all public APIs for this task. I extensively used GPT-3 for it, and that's my professional judgement.

As for AI validation, it can be 10x more comfortable than manual information extraction, but still requires about 50% of the manual effort. It is not making people suddenly 10x more effective.

Not even OCR is 100% accurate. The best systems have 95% accuracy on noisy document scans. One digit or comma could make the whole transaction absurd, if you send those money without checking you could go in bankruptcy.

The best models we have today are good at generating correct answers 90% of the time - code, factual questions, reasoning. They can do it all but not perfectly. We don't know the risks and can't use this level of confidence without human in the loop.

13

visarga t1_j763vpj wrote

Reply to comment by Neurogence in Possible first look at GPT-4 by tk854

That depends a lot on context window size - if it's 4K or 8K tokens like today, it won't cut it. For full app coding you need to be able to load dozens of files.

Related to this - if we get, say... 100K context size, we could just "talk to books" or "talk to scientific papers".

5

visarga t1_j744swb wrote

If you put Stable Diffusion or chatGPT to generate automatically without human review or prompting, they will generate tons of garbage. Generative AIs are garbage until someone stamps their work as good. So they need humans to be worth anything. They are just junk on their own. It's a long way off from job replacement - even self driving cars require human at the wheel. These AIs still hallucinate facts, who can use them as they are now. Clearly someone will have to find a way before they can get useful without being babysitted.

1

visarga t1_j741kg5 wrote

> AI will dominate human labor and push them out of the market.

AI teamed with a human will dominate both AI and human. AI is much better with a human, and humans are better with AI. Since we have competition, every company will have to add AI to their current workforce and keep the people, because they are the differentiating factor. You can scale AI in the cloud, but you can't simply spawn people.

1