Comments

You must log in or register to comment.

TheLastSamurai OP t1_it68lrg wrote

Summary

We expect that there will be substantial progress in AI in the next few decades, potentially even to the point where machines come to outperform humans in many, if not all, tasks. This could have enormous benefits, helping to solve currently intractable global problems, but could also pose severe risks. These risks could arise accidentally (for example, if we don’t find technical solutions to concerns about the safety of AI systems), or deliberately (for example, if AI systems worsen geopolitical conflict). We think more work needs to be done to reduce these risks.

Some of these risks from advanced AI could be existential — meaning they could cause human extinction, or an equally permanent and severe disempowerment of humanity.2 There have not yet been any satisfying answers to concerns — discussed below — about how this rapidly approaching, transformative technology can be safely developed and integrated into our society. Finding answers to these concerns is very neglected, and may well be tractable. We estimate that there are around 300 people worldwide working directly on this.3 As a result, the possibility of AI-related catastrophe may be the world’s most pressing problem — and the best thing to work on for those who are well-placed to contribute....

3

artix111 t1_it6osyf wrote

I am pretty sure it will be revolutionary in this decade already and that things will change faster than we can comprehend.

1

Endward22 t1_itcy4vs wrote

I don't see why everyone always assumes that a potential AI wants to kill us.

I rather think that the AI is just so far in the future that we can't imagine it.

1

TheLastSamurai OP t1_ited1ji wrote

You should read the article. Some very smart people but the chance at 10%, also look at AI scaling, it follows the concept of exponential growth, it will explode in capabilities so fast once it gets going.

1

prototyperspective t1_it6wneb wrote

Info like that should be contained in articles like https://en.wikipedia.org/wiki/Existential_risk_from_artificial_general_intelligence https://en.wikipedia.org/wiki/Regulation_of_algorithms (and that page may be useful in expanding these albeit the summaries should be shorter than that page).

There could also be a new article but I think the "Existential risk" article is too specific to "Existential risk" which makes the common mistake of ignoring the more immediate currently more realistic issues.

Also often it may be a better approach to explore regulating potential AI-made products such as regulation of chemicals (or similar fundamentally different approaches).

2

Endward22 t1_itcy7qw wrote

>"Existential risk"

I heart this from Borstrom, tbh.

1

prototyperspective t1_itd1ubv wrote

I'm not saying there is no existential risk but it's only a part of major risks and it's not a good approach to attempt to disentangle these or to ignore other ones so it should be an article about risks with info that some of them may be existential with some info why/how so.

2

FuturologyBot t1_it6bz56 wrote

The following submission statement was provided by /u/TheLastSamurai:


Summary

We expect that there will be substantial progress in AI in the next few decades, potentially even to the point where machines come to outperform humans in many, if not all, tasks. This could have enormous benefits, helping to solve currently intractable global problems, but could also pose severe risks. These risks could arise accidentally (for example, if we don’t find technical solutions to concerns about the safety of AI systems), or deliberately (for example, if AI systems worsen geopolitical conflict). We think more work needs to be done to reduce these risks.

Some of these risks from advanced AI could be existential — meaning they could cause human extinction, or an equally permanent and severe disempowerment of humanity.2 There have not yet been any satisfying answers to concerns — discussed below — about how this rapidly approaching, transformative technology can be safely developed and integrated into our society. Finding answers to these concerns is very neglected, and may well be tractable. We estimate that there are around 300 people worldwide working directly on this.3 As a result, the possibility of AI-related catastrophe may be the world’s most pressing problem — and the best thing to work on for those who are well-placed to contribute....


Please reply to OP's comment here: https://old.reddit.com/r/Futurology/comments/y9lj2u/preventing_an_airelated_catastrophe_problem/it68lrg/

1