Viewing a single comment thread. View all comments

visarga t1_is0fpyb wrote

I became aware of AI in 2007 when Hinton came out with Restricted Boltzmann Machines (RBMs, a dead end today). I've been following it and started learning ML in 2010. I am a ML engineer now, and I read lots of papers every day.

Ok, so my evaluation - I am surprised with the current batch of text and image generators. The game playing agents and the protein folding stuff are also impressive. I didn't expect any of them even though I was following closely. Two other surprises along the way were residual networks, which put the deep into deep learning, and the impact of scaling up to billions of parameters.

I think we still need 10,000x scaling to reach human level both in intelligence and efficiency, but we'll have expensive to use AGI in a lab sooner.

I predict the next big thing will be large video models, not the ones we see today but really large like GPT-3. They will be great for robotics and automation, games and of course video generation. They have "procedural" knowledge - how we do things step by step - that is missing in text and images. They align video/images with audio and language. Unfortunately videos are very long, so hard to train on.

3

mealoftheday42 t1_is10epb wrote

What would your guess be for how much we need to scale for expensive agi? Are we talking 50x, 500x, 1000x, 5000x?

Feel free to ignore this rant (I mostly wrote it to get my own thoughts in order), but if you have time by all means correct me on anything I'm wrong about.

I know this is largely guesswork as to the needed model scale/compute power, but I'm a layman who generally disregarded the possibility of near-term agi... until I saw a few arguments showing the freakiness of the exponential curves at play. This has made me feel a bit existential to say the least. I know confirmation bias is a thing so I've sought out counterarguments. From what I've seen they mostly center around two points. The first is that long-term gains with our current pathway are untenable (our current ai lacks any sort of true intelligence and will quickly plateau), and that failed examples like self-driving cars are proof of how ill-equipped we are to deal with such complex problems.

To the dismay of my existentialism, neither of these arguments convince me. To the first point, the nay-sayers have a habit of bringing less data to their arguments than the near-termers. I don't doubt another ai winter is possible, but a stagnation first requires a slowing in progress. Exponential curves don't just slam to a halt, barring global catastrophe. The only place I see evidence of stagnation is in Moore's law, which is a pretty narrow measure. Even then, computing power by other measures is still expanding just as fast, and most engineers I've heard from still expect a few more doublings on the transistor counts. I don't doubt that just brute forcing agi by bringing more computing power to our current frameworks is unrealistic, but that's when I look at other measures of ai progress. GPUs produced, spending on ai, number of people working on ml; they're all expanding exponentially. People in the 60s famously predicted famine would engulf the rising population, but failed to anticipate that rising population would bring additional innovation. Hunger has only gone down at an exponential rate since then. Expecting engineers to suddenly stop looking for new ml frameworks is to ignore all historical trends.

To the second point, I agree that people have been claiming self-driving cars are two years away for the last ten years. This to me though seems more anecdotal than anything. Sure, a problem that optimists said would be solved in five to ten years is likely going to take twenty, but a ten year error is a blink of an eye all things considered. When you broaden the range of exponential technological progress to the start of the industrial revolution, this delay seems to be a molehill rather than a mountain. Regardless, we can find plenty of counter-examples of people underestimating the time it'd take for ai to do things like beat humans at Go. When both sides are just using anecdotes, only data can resolve whether we're systematically overestimating or underestimating progress. Unfortunately I've been unable to find anything on this, so for now I'm unconvinced by the argument we're too optimistic. I'm keeping an eye on prediction markets in the future though.

1

Frumpagumpus t1_is5bmlr wrote

i'm gonna answer for him and say, just put all the points you mentioned on a normal/gaussian distribution with 500x and 1000x one std deviation away from mean, and 50x and 5000x two std deviations

that's what I guess he would think

2

TopicRepulsive7936 t1_is6d7q6 wrote

I think people tend to vastly underestimate what we're capable of in terms of computing infrastructure. In the 1930's AI was done on pen and paper (Turing's chess engine) and in the early 1990's supercomputers were still a single cabinet systems. We're far removed from those times but there's also room to grow. Then there are those who have kneejerk opposition of us utilizing our planetary resources for AI, they are entitled to their opinion but I don't think it does any good to listen to those types of people.

1