Viewing a single comment thread. View all comments

Codydw12 t1_jcdswpf wrote

> In a 2022 survey, A.I. experts were asked, “What probability do you put on human inability to control future advanced A.I. systems causing human extinction or similarly permanent and severe disempowerment of the human species?” The median reply was 10 percent.

> I find that hard to fathom, even though I have spoken to many who put that probability even higher. Would you work on a technology you thought had a 10 percent chance of wiping out humanity?

> We typically reach for science fiction stories when thinking about A.I. I’ve come to believe the apt metaphors lurk in fantasy novels and occult texts. As my colleague Ross Douthat wrote, this is an act of summoning. The coders casting these spells have no idea what will stumble through the portal. What is oddest, in my conversations with them, is that they speak of this freely. These are not naifs who believe their call can be heard only by angels. They believe they might summon demons. They are calling anyway.

> I often ask them the same question: If you think calamity so possible, why do this at all? Different people have different things to say, but after a few pushes, I find they often answer from something that sounds like the A.I.’s perspective. Many — not all, but enough that I feel comfortable in this characterization — feel that they have a responsibility to usher this new form of intelligence into the world.

> A tempting thought, at this moment, might be: These people are nuts. That has often been my response. Perhaps being too close to this technology leads to a loss of perspective. This was true among cryptocurrency enthusiasts in recent years. The claims they made about how blockchains would revolutionize everything from money to governance to trust to dating never made much sense. But they were believed most fervently by those closest to the code.

So throw it all in the trash? Stop fighting demons? Or is it worth it to take a risk that we might burn out in an attempt to create technologies that progress to the point of immense benefit? This just reads like fearmongering.

I do not see AI as some cure all nor do I believe it will completely replace humanity as some on here seem to believe, but I do believe that a lot of the benefits that could come from it are worth it.

> Could A.I. put millions out of work? Automation already has, again and again. Could it help terrorists or antagonistic states develop lethal weapons and crippling cyberattacks? These systems will already offer guidance on building biological weapons if you ask them cleverly enough. Could it end up controlling critical social processes or public infrastructure in ways we don’t understand and may not like? A.I. is already being used for predictive policing and judicial sentencing.

Again, fearmognering. Automation and job loss is a constant fear. Terrorists and bad actors are always feared to get nukes yet none have to date. The predictions can also help eliminate disease and help crime prevention by helping those in need who are often the most predisposed to commit crime.

−3

Coachtzu t1_jcdxi2h wrote

You're cherry picking. He addresses this in the article. We can't afford to be left behind, yet we also don't understand what we are racing towards.

Automation has also already cost jobs. It will cost more. This is not controversial. We need to figure out how we adapt to a world where our work does not and should not define us.

20

iStoleTheHobo t1_jce0127 wrote

>We need to figure out how we adapt to a world where our work does not and should not define us.

Precisely. Nobody seems to talk about this particular point but let's put it like this: If the artificial intelligence revolution will be bigger than the splitting of the atom why the hell would we allow the private sector to govern these tools? Do we allow private companies to handle atom bombs?

15

Cheapskate-DM t1_jce1kn8 wrote

Atom bombs require uranium. Uranium comes from mines. Mines occupy land. And if governance has any talent which it can reliably manage, it's keeping people away from a given piece of land.

Code has no such restriction.

5

Codydw12 t1_jcf9wnq wrote

> You're cherry picking. He addresses this in the article. We can't afford to be left behind, yet we also don't understand what we are racing towards.

> > But I don’t think these laundry lists of the obvious do much to prepare us. We can plan for what we can predict (though it is telling that, for the most part, we haven’t). What’s coming will be weirder. I use that term here in a specific way. In his book “High Weirdness,” Erik Davis, the historian of Californian counterculture, describes weird things as “anomalous — they deviate from the norms of informed expectation and challenge established explanations, sometimes quite radically.” That is the world we’re building.

> > I cannot emphasize this enough: We do not understand these systems, and it’s not clear we even can. I don’t mean that we cannot offer a high-level account of the basic functions: These are typically probabilistic algorithms trained on digital information that make predictions about the next word in a sentence, or an image in a sequence, or some other relationship between abstractions that it can statistically model. But zoom into specifics and the picture dissolves into computational static.

> > That is perhaps the weirdest thing about what we are building: The “thinking,” for lack of a better word, is utterly inhuman, but we have trained it to present as deeply human. And the more inhuman the systems get — the more billions of connections they draw and layers and parameters and nodes and computing power they acquire — the more human they seem to us.

None of this seems actually profound or useful to me. Saying that the AIs that we build will be alien to our own thinking? To me that, in his own words, is in the laundry list of obvious.

> Automation has also already cost jobs. It will cost more. This is not controversial. We need to figure out how we adapt to a world where our work does not and should not define us.

And that I fully agree with but every time I suggest heavily taxing automated jobs as a means to fund Universal Basic Income I have hypercapitalists call me a socialist for believing people should be allowed to live without the need of working.

5

Coachtzu t1_jcfatna wrote

>None of this seems actually profound or useful to me. Saying that the AIs that we build will be alien to our own thinking? To me that, in his own words, is in the laundry list of obvious.

I don't know if I think it's profound either, but I do think it's a healthy reminder. Its a good reminder that we don't really understand these algorithms, and that regardless of how human-presenting they are, they are not human and we can't trust them to act in certain ways. Maybe not particularly helpful, but worthwhile none the less (in my opinion).

>And that I fully agree with but every time I suggest heavily taxing automated jobs as a means to fund Universal Basic Income I have hypercapitalists call me a socialist for believing people should be allowed to live without the need of working.

This has happened to me too, I've suggested exactly the same thing (though admittedly stole the idea from mark Cuban when he guest hosted on a podcast at one point). At this point everything is socialist if it's different than the status quo though so I try to ignore it.

2

Codydw12 t1_jcfhsmv wrote

>I don't know if I think it's profound either, but I do think it's a healthy reminder. Its a good reminder that we don't really understand these algorithms, and that regardless of how human-presenting they are, they are not human and we can't trust them to act in certain ways. Maybe not particularly helpful, but worthwhile none the less (in my opinion).

And this is fair. AI will not act like a human nor will it be completely logical in every aspect. We don't actually know how one will act or react or what its been trained on.

> This has happened to me too, I've suggested exactly the same thing (though admittedly stole the idea from mark Cuban when he guest hosted on a podcast at one point). At this point everything is socialist if it's different than the status quo though so I try to ignore it.

Indeed. I have given up on trying to predict future economies but the current system won't work much longer.

2

Coachtzu t1_jcfpowm wrote

100% agree. Appreciate the good discourse, it's hard to find on here.

3

Lettuphant t1_jcfdtv1 wrote

This reminds me of two early examples of evolutionary AI design (though I doubt I could find the details, these are from interviews long ago: one was a circuit board that an AI designed which looked non-fictional and that no human would have designed, but it worked. Best guess was the EM of one part was interacting with another.

The other is some guys who were trying to build the lightest possible body for a drone, and set an AI to building it. They 3D printed it and when their buddy who was a veterinarian walked in he said "hey, why do you have a flying squirrel skeleton here?" AI doing what natural selection took millions of years to, running through the iterations in milliseconds rather than generations.

2