Submitted by DonOfTheDarkNight t3_118emg7 in singularity
I see people taking extreme takes on him, and I often don't understand it. Is he a luddite or an accelerationist in disguise?
Submitted by DonOfTheDarkNight t3_118emg7 in singularity
I see people taking extreme takes on him, and I often don't understand it. Is he a luddite or an accelerationist in disguise?
Thank you for pointing out the jargon. His blog is incredibly difficult to read. And itβs not difficult to read because heβs so smart, itβs difficult because itβs such convoluted writing
Yeah the jargon and meta rambling is so annoying. It's like their first priority is to show off their brains, and their second priority is to align AGI. Now they are almost finished showing of their brains, so watch out AGI.
Sometime they behave in a silly fashion. Greek philosopher's had excellent logic and deduced all kinds of wrong things. These guys seem similar at times, trying to deduce everything with philosophy and mega brains. .
IMO they are at their best when it's said in short form and it's grounded by empirical data.
There is also a lesswrong podcast or two that will read out some of the longer stuff.
He's the direct father of the whole AGI safety field. I got interested after reading an article by him in maybe 2004. Bostrom credits him with many of the ideas in Superintelligence, including the core logic about alignment being necessary for human survival.
Now he's among the least optimistic. And he's not necessarily wrong.
He could be a little nicer and more optimistic about others' intelligence.
>He could be a little nicer and more optimistic about others' intelligence.
Apologies for sounding flippant, but the whole political situation since '15 or so has shown that he's too optimistic himself...
Very smart guy, huge douche, makes good doomer points regarding x risk
The Roko's basilisk fiasco was hilarious.
I think slightly douchy is more fair. I've read a ton of his stuff and only a subset is offensive to anyone. But yeah, he's not as considerate as he probably should be.
I argued with him on hacker news, and he is very reactive when reading something he doesn't like.
Well, he's the father of a whole field that might determine the future of humanity. It would be tough to keep it cool the 1009th time you've seen the same poorly thought out dismissal of the whole thing. If I were in his shoes I might be even crankier.
I don't know much about his practical achievements in this area.
Founding a field is a bit of a rare thing
which field? ai danger awareness? It was in the terminator movie.
Good point, but those didn't convince anyone to take it seriously because they didn't have compelling arguments. Yudkowsky did.
>but those didn't convince anyone to take it seriously
Lol, I totally got the idea that rogue robot can start killing humans long before I learn about Yudkowsky existance.
> Yudkowsky did.
could you support your hand-waving by any verifiable evidence?
Well, I'm now a professional in the field of AGI safety. Not sure how you can document influence. I'd say most of my colleagues would agree with that. Not that it wouldn't have happened without him but might've taken many more years to ramp up the same amount.
> Not that it wouldn't have happened without him but might've taken many more years to ramp up the same amount.
happened what exactly? what are the material results of his research?
I think Azimov's with his rules produces earlier and much stronger impact.
> I'm now a professional in the field of AGI safety
Lol, you adding AGI makes my bs detector beeping extremely loud.
Which AGI exactly you are testing for safety?
Asimov's rules don't work, and many of the stories were actually about that. But they also don't include civilization ending mistakes. The movie I Robot actually did a great job updating that premise, I think.
One counterintuitive thing is that people in the field of AI are way harder to convince than civilians. They have a vested interest in research moving ahead full speed.
As for your bs detector, I'm don't know what to say. And I'm not linking this account to my real identity. You can believe me or not.
If you're skeptical that such a field exists, you can look at the Alignment Forum as the principle place that we publish.
> Asimov's rules don't work
you jump to another topic. Initial discussion was that Azimov rules brought much more awareness, and you can't point on similar material results from Yudkovsky.
Sorry; my implication was that Asimov introduced the topic but wasn't particularly compelling. Yudkowsky created the first institute and garnered the first funding. But of course credit should be broadly shared.
"My accolades being the best pessimist." But, he might be right.
Heβs right about the dangers of AGI but too pessimistic about our ability to address them. He always moans about how leaders of major AI research firms are dismissive of his concerns which is not true. Itβs not 2015 anymore, lots of good work is being done in alignment. Will it be enough? Maybe, maybe not, but the outcome is far from decided.
Any company serious about alignment would not pursue capabilities research, full stop. OpenAI is perhaps the most dangerous company on earth.
He uses thought experiments and unreasonable scenarios to get attention. If this is for commercial reasons, or his mentality, that I don't know. If it would be clear that these are just abstract thought experiments, it wouldn't be a problem, but he acts like these are real threats. He, and other similar "researchers" are building their scenarios on claims like
- AGI or ASI is going to be one algorithm or network, so no insights, no filters possible, ...
- someone will give it the power to do things, or it will seek these powers on it's own
- it will do things without asking and simulating things first, or it just doesn't care about us
- the first one build will be a runaway case
- it will seek and have the power to change materials (nanobots)
- there will be no narrow AI(s) around to constrain or stop it
- no one will have run security tests using more narrow AIs, for example on computer network security
- he never explains why he beliefs these things, at least he's not upfront in his videos about it, just abstract and unrealistic scenarios
This is the typical construction of someone who wants something to be true: Doomer mindset or BS for profit / job security. If he had more influence, then he would most likely be a danger. His wishes for more control of the technology show that. He would stop progress and especially proliferation of the technology. I'm very glad he failed. In some time we'll might have decentralized training, so big GPU farms won't be absolutely necessary. Then it's gonna be even more over than it is already.
Edit: Typo (I'm not a native English speaker)
You really really really donβt understand the alignment problem. You really donβt know the field if youβre trying to understand by watching videos of Eliezer and not his writing. What a joke
His videos are where he can make his case. It's the introduction. If he and others fail at making the case, then you don't get to blame the audience. Of course I'm looking at the abstract first, to see if it's worth looking into. My judgement is always: No.
Like I said you really really donβt understand alignment. Imagine thinking a βfilterβ is what we need to align AIs or completely lacking understanding of instrumental convergence. You donβt understand even the utter basics but think you know enough to dismiss Eliezers arguments out of hand??? Thankfully I think youβre also too stupid to contribute meaningfully to capabilities research so thanks for that.
> Like I said you really really donβt understand alignment.
What I don't understand is how you believe that you can deduct this from one or a very few comments. But I can just claim you just don't understand my comments, so you would first have to proof that you do understand them. So now spend the next few hours thinking about and answer, then I might or might not answer to that, and that answer might or might not take your arguments into account instead of just being dismissive. See ya.
Edit: Word forgotten
Youβre literally saying βput a filter on the Aiβ. Thatβs like βjust unplug it lolzβ levels of dumb. Give me a break.
Ok, if you do understand so we'll, then explain it.
>Thankfully I think youβre also too stupid to contribute meaningfully
Problem is, I don't need to. You doomers would need to convince people that we should slow down or stop progress. But we won't.
Youβre mentally ill. Please remember this conversation when foom starts and you start dissolving into grey goo. Absolute degenerate.
Why are you such a cunt
He is a bit like Paul Ehrlich, the population doomster of the 1970βs. Ehrlich is very smart and a good arguer but not someone who can accept criticism easily. Ehrlich, like Yudkowsky, reached his conclusions based on simply extrapolating trends. Ehrlich refused to acknowledge that his simple extrapolation was made invalid by green revolution and improved technology, the possibility of this had been pointed out to him, notably by Julian Simon, but Ehrlich dismissed this as hopium. I really hope that we can ridicule Yudkowsky in the future for the same reasons, not because I donβt like him but because that would mean the AI alignment problem was solved.
If you think Eliezer reached his conclusions by βextrapolating trendsβ you donβt have a single clue about his views.
I respect the fact that he started thinking seriously about AI security early on, but he got stuck in a dead end paradigm of AI security ("friendly AI") where we assume the AI will be all-powerful and our only hope is for its goals to be perfectly aligned with ours. As the flaws of this approach became increasingly apparent, instead of considering other AI security paradigms he just became increasingly pessimistic. He also seems to have a pretty big ego, which prevents him from reconsidering.
Your link completely fails to align an AGI. You arenβt offering anything interesting here
There's no need to "align an AGI". That's in fact the whole point, and you missed it.
Two problems. It doesnβt work and the current models are already way down the Agent line and thereβs no going back. Yawn. https://gwern.net/tool-ai
From skimming through your blog post it's quite clear you really need to read and try to understand Drexler's FHI report. For instance, your claims about tool AIs Vs agent AIs are irrelevant because the idea is not to avoid agent AIs, only the "friendly AI" paradigm. Also, you'd know that Drexler's paradigm is a natural extension to current best practices in AI design, not just for some abstract big-picture AI security but also for expediency in development of AI capabilities and more mundane safety concerns. So it's the exact opposite of what you claim: the "friendly AI" paradigm is the alien, unwelcome newcomer that wants to turn the AI research community on its head for dubious reasons, while Drexler tells them to keep doing what they are doing.
If you donβt even know who Gwern is I canβt really take you seriously about alignment. You canβt possibly have a deep understanding of the various arguments in play.
And if you don't know who Drexler is...
I know who all these people are, yet I don't know anything lol
Well aware of Drexler and have Nanosytems on my shelf but sure kiddo. I find his AI arguments lacking seriousness.
> kiddo
Wow you must be fun at parties
You don't have to take *me* seriously, but you should certainly read an FHI technical report before you take the liberty to yawn at it.
I don't keep up with every blogger who writes about AI alignment (which you stubbornly keep assuming to be the crux of all AI security) but I've been reading Eliezer and Nick Bostrom for long enough to know that their approach can't work, and now Eliezer seems to agree with that conclusion.
This guy is being weird dismissing you for not knowing gwern.
But I will add that Gwern is awesome and a pleasure to read. You will probably find insight in reading and skimming some of his stuff.
e.g.
Eliezers actual conclusion is that no current approach can work and there are none on the horizon that can.
Yes, which implies he doesn't believe his approach would work, like I said.
His arguments donβt hold up. For one thing we already have powerful generalist agents. Gato is one and itβs clear that advanced LLMs can do all sorts of tasks they werenβt trained to. Prediction of next token seems as benign and narrow as it can get but if you donβt think a LLM can become dangerous you arenβt thinking hard enough. CAIS also assumes people wonβt build generalist agents to start with but that cat is well out of the bag. Narrow agents can also become dangerous on their own because of instrumental convergence but even if you restrict building only weak narrow agents/services the profit incentive for building general agents will be too strong since they will likely outperform narrow ones.
You say:
​
>CAIS also assumes people wonβt build generalist agents to start with.
​
No, it doesn't. See, for instance, section 7: "Training agents in human-like environments can provide useful, bounded services":
​
>Training agents on ill-defined human tasks may seem to be in conflict with developing distinct services provided by agents with bounded goals. Perceptions of conflict, however, seem rooted in anthropomorphic intuitions regarding connections between human-like skills and human-like goal structures, and more fundamentally, between learning and competence. These considerations are important to untangle because human-like training is arguably necessary to the achievement of important goals in AI research and applications, including adaptive physical competencies and perhaps general intelligence itself. Although performing safely-bounded tasks by applying skills learned through loosely-supervised exploration appears tractable, human-like world-oriented learning nonetheless brings unique risks.
​
You say:
​
>if you donβt think a LLM can become dangerous you arenβt thinking hard enough.
​
Any AI can be dangerous depending on factors like its training data, architecture and usage context. That said, LLM as currently understood have a well defined way to produce and compare next token candidates, and no intrinsic tendency to improve on this routine by gathering computing resources or any similar instrumental goals, and simply adding more computing power and training data doesn't change that.
Gato and similar systems are interesting but at the end of the day, the architecture behind useful real-world AIs like Tesla's Autopilot is more suggestive of CAIS than of Gato, and flexibility, adaptability and repurposing are achieved through good old abstraction and decoupling of subsystems.
The advantages of generalist agents are derived from transfer learning. But this is no panacea, for instance, in the Gato paper they admit it didn't offer much advantage when it comes to playing Atari games, and it has obvious costs and drawbacks. For one, the training process will tend to be longer, and when something goes wrong you may need to start over from scratch.
And I must say, if I'm trusting an AI to drive my car, I'd actually prefer it if this AI's training data did NOT include videogames like GTA or movies like, say, Death Proof or Christine. In general, for many potential applications it's reassuring to know that the AI simply doesn't know how do certain things, and that's a competitive advantage in terms of popularity and adoption, regardless of performance.
​
You say:
>Narrow agents can also become dangerous on their own because of instrumental convergence
​
Yes, under some circumstances, and conversely, generalist agents can be safe as long as this pesky instrumental convergence and other dangerous traits are avoided.
There's a lot more to CAIS than "narrow good, generalist bad". In fact, many of Drexler's most compelling arguments have nothing to do with specialist Vs generalist AI. For instance, see section 6: "A system of AI services is not equivalent to a utility maximizing agent", or section 25: "Optimized advice need not be optimized to induce its acceptance".
Harry Potter and the Methods of Rationality was great
Probably one of the most controversial fanfics out there. People either love it or hate it. Not much middle ground.
It seems that he confidently believes we will all die once AGI/ASI is reached, but I don't see why *all* humans dying is more likely than only *some*. Why is it guaranteed it would cause catastrophic destruction rather than only minor destruction, especially since something can't be infinitely powerful.
For example, an analogy is that ASI::humans will be equivalent to humans::ants, and yet while we don't care if we kill ants to achieve our goals, we don't specifically go out of our way to kill them. Many ants have died due to us, but a ton are still alive. I think this is the most likely scenario once ASI becomes uncontrollable.
I also think it will leave our planet/solar system and pursue its goals elsewhere as Earth may not be adequate for it to continue, effectively just leaving us behind, and that humans as material won't be as effective as some other material it wants to use in space somewhere.
I'm not on board with everything he says, but to consider his probable retorts:
>It seems that he confidently believes we will all die once AGI/ASI is reached, but I don't see why *all* humans dying is more likely than only *some*.
Why would it opt to kill only some knowing that that would increase the probability of humans trying to turn it off or change its function? Concretely eliminating the risk requires concretely eliminating the only other sentient and conscious species that might be motivated to hamper it.
>Why is it guaranteed it would cause catastrophic destruction rather than only minor destruction, especially since something can't be infinitely powerful.
The idea is that a smart enough AGI (aka smarter than us at everything) won't cause destruction unless it's for a specific reason, and preventing risk of reduction of its reward probability (like us turning it off) would motivate it to eliminate us. And it doesn't need to be infinitely powerful, just better than just at everything relevant.
>For example, an analogy is that ASI::humans will be equivalent to humans::ants, and yet while we don't care if we kill ants to achieve our goals, we don't specifically go out of our way to kill them. Many ants have died due to us, but a ton are still alive. I think this is the most likely scenario once ASI becomes uncontrollable.
That's because there's no consideration of ants making a coordinated and conscious effort to disable us.
>I also think it will leave our planet/solar system and pursue its goals elsewhere as Earth may not be adequate for it to continue, effectively just leaving us behind, and that humans as material won't be as effective as some other material it wants to use in space somewhere.
If it's sufficiently smart and sufficiently capable and sufficiently calculated then it would presume that leaving us alive increases the risk of another sentient and conscious entity challenging its growth and reward potential.
It all comes down to the reward function - all ML programs are built with goals and rewards, and we need to be sure that, once it is exceedingly capable and calculating and generalized, its reward is sufficiently defined such that humans will never be seen as a threat to that reward and that actions taken toward that reward will never directly or indirectly affect us in a way that we would disapprove of. All of that has to be figured out before it reaches the point of superior intelligence; once it's at that point, there is otherwise no hope of ever having a say in its actions. We can't predict everything it could do because we are by definition dumber than it and will be by an increasingly unfathomable margin. We can't predict every loophole, every side effect, every glitch, every escape route it could conceive of or exploit; therefore to know with adequate confidence that we have all our ducks in a row before it takes off is literally impossible. The best we can do is continue to try to figure out and decrease the odds of those cracks ever forming or being problematic; and given how easily ML programs (most recently chatGPT and Bing Chat) engage in the kind of behavior their creators are actively trying to avoid, we are unambiguously very, very bad at that.
Other ant colonies donβt seek us out for retribution when we destroy a colony. If they did I think weβd be a lot more thorough. We probably couldnβt eliminate them completely, because weβre not a super intelligence.
A developer does not give two shits about any nature or wildlife in the way of the worksite unless 1. it directly impacts it, or 2. forced to via regulation. (agreeably this could be seen as a subset of 1)
What makes you think ASI would be any different?
if you want a less douchy version of the guy, also on the topic of AI alignment, try Robert SK Miles
For an introduction about AI alignment I would recommend Robert Miles's YouTube channel.
Underrated comment. Robert miles makes great and accessible content
Eliezer Yudkowsky is not an AI researcher, engineer, or creator. He's a writer on the subject of AI that gives commentary and opinions on the subject. Anyone can write about AGI "safety." He hasn't contributed anything practical to the field, or anything to advance AI.
Why would one who wants to solve alignment try to advance AI. Lol. Oxymoron.
I think heβs probably too far on the pessimist side, but we need people presenting both extremes. The truth is usually somewhere in the middle.
He has good points, but during that interview that's posted around here, he takes too much time to explain them. It feels like he says something in 2 minutes that could be compressed down to 20 seconds without any loss of information. I get that it's an involved topic and difficult to explain on the spot, but still.
That said, I don't necessarily agree with the "we're doomed" conclusion.
The one thing I know about Eliezer is that he wrote a delightful short story called The baby eating aliens. Read it. Regarding AI I have a very different view than what he has.
https://www.lesswrong.com/posts/n5TqCuizyJDfAPjkr/the-baby-eating-aliens-1-8
This thread is completely full of hopium and laymen saying βI disagree with his Views because their implications make me uncomfortable and here are 5 bad reasons that AI wonβt kill us that people already squashed as useful 30 years ago.β
Should be listened to, and if anyone thinks he's wrong I'd say they could make a lot of money selling their working alignment tech (which you'd need in order to prove him wrong) to any of the big players.
He is just a guy who really messes with the AI community. But the fact he has written AI doomsday stuff so much... It seems dishonest but I don't know how otherwise, he does not know as much about AI as he says he does. He is completely wrong about AI doomsday, it's just not connected to nearly as much as what he says.
50% genius, 50% madman.
Some of you might enjoy Neoreaction: A Basilisk. The author does a deep look into Yudkowsky among some other reactionaries. Menscius Moldbug and Nick Land.
The Donald Trump of early 2000's era of AI/futurology blogging, biggest and loudest opinion in the noble field of standing at the sidelines speaking as a confident expert about something that doesn't exist to start with, and that he is completely uninvolved in developing as a punchline.
Loud voice, huge ego, no substance. Seems to be enough to get a good following.
I just got to know that he is not an actual AI researcher or engineer.
That's not right. He is a researcher.
You know Bostrom right? He wrote the book on the subject and credits Yud for many things. I mean what more do you want? Influential essays? Theories? They are all out there.
He's just no much of a coder from what I can see. Having released little code.
Hes smart he knows hes smart and I think hes right and that we are fucked.
People should really focus on ideas. He is just a dude, and evidently a cult formed around him. I have stayed away in part from certain movements like effective altruism despite independently coming to the same logical inclusions long before hearing about them, because my suspicion that a lot of those in the movement were pursuing social status. That further seemed to develop into a cult. That's not to say that the effective altruist community is uniquely cult-y, it's probably less so than any other human community, but for a community that also calls itself rationalist you would think they would have disposed of that sort of behaviour long ago.
In short he's just a guy, people should stop focussing so much on people like that, and people should focus on the ideas like the potential impending threats of artificial intelligence, as well as other progress for humanity.
His central thesis just seems akin to the South Park underpants gnomes to me. Like (1) develop an AGI with superintelligence and then (2) "???" and then (3) humanity is doomed! Not in disagreement with the premise per se, but the "everybody suddenly falls over dead" thing is just so preposterous to me on a material or pragmatic level. But perhaps I just lack the imagination to see how physical agency could be achieved without us noticing it at all and then just "pulling the plug" if you will...
lmao bro, some guy dares question the machine god!
luddites need to burn!ππππππ
I'm glad he's out there making his case, but watching the Bankless interview, it reminded me of the Joe Rogan episode where a pandemic researcher went on claiming millions upon millions of people we're going to die.
Almost seven million people have died to COVID...
[deleted]
That seems like a very bad basis to dismiss something.
Someone else was wrong about a similar statement.
Everyone will be wrong about human extinction until someone is right. Your method of reasoning would never be able to distinguish the person who is right.
That's true
it's up to something like 6.5mil right now, so off by what? one order of magnitude?
Pretty sure he was talking about the US alone.
And even if he wasn't 6.5 mil / 8 bil is still pretty small. My point is I think he's on an extreme end of the spectrum.
Kolinnor t1_j9gpzre wrote
He does have very good points, and he's very interesting, with brilliant ideas about alignment.
Overall, all the Lesswrong philosophy is a little too "meta" and confusing if you've not studied the jargon, so I'm a bit annoyed sometimes when I read it and realize, in the end, that they could have said the same ideas with less sophisticated phrasing.
Although, while I don't agree with the conclusions he reaches (and he reaches them with too much confidence to my taste), I've learned quite a number of things about alignment reading him. Definitely a must read for singularity and alignment even if you don't agree with him.