WarImportant9685

WarImportant9685 t1_j1nfl62 wrote

Frankly, I don't know. Cause nuclear is a grey area tho. In one hand it cause peace on modern times. On the other hand there is always the threat of extinction by nuclear winter.

But let's leave nuclear for a second. I can give you an example of a technological progress where humanity dropped the ball. Usage of CFC refrigerant which damages ozone layer. While it's cool that humanity have banned CFC refrigerant as of now, it takes 20 years for the treaties that banned it (Montreal protocol) to be created from the discovery of the problem.

My point is, humanity is great at advancing tech, but maybe not so much at advancing tech safely. I think at least people should take seriously that aligned AGI is not a guaranteed matter. And instead of getting defensive, ideally more people would want to help alignment research to make sure unaligned AGI not be deployed.

1

WarImportant9685 t1_j1lr2tx wrote

I think the correct view is that each technology should be evaluated on its own. And cannot be blanket viewed as good. I'd agree with technology in general is good. But I'd disagree with claims along the line of 'all technological progress is good'. I understand that most people are looking at heuristic reasons why technological progress is good. And I would agree that humanity history is the history of technological progress.

But it's dangerous to say that AGI is good because technology progress is good.

And I kinda disagree with your comments about nuclear weapon. Fact is we are lucky that USA get nuclear weapon first instead of germany. I viewed nuclear weapon as force multiplier instead of good or evil.

And I viewed unaligned AGI as evil. While aligned AGI as force multiplier.

More intuitively, throughout history, we mostly got technology that is still smaller than life. Fire, wheels, masonry, etc. The issue is that right now our tech have started to become larger than life, starting with nuclear and likely larger than the whole humanity with AGI.

1

WarImportant9685 t1_j1l4lx2 wrote

You are missing some biases that's on the optimistic camp brother

  • Techno-religiousity -> Simple believe that any progress on technology is ultimately good.
  • Unrealistic optimism -> Some might call this hopium
  • Intelligence anthromorphism -> Believe that intelligent agent would be similar to human
1

WarImportant9685 t1_j173veq wrote

Okay, then imagine this. In the future an AGI is in training to obey human being. In the training simulation, he is trained to get groceries. After some iterations, where unethical stuff happens (robbery for example), he finally succeed to buy groceries as human wanted it.

Question is, how can we be sure that he isn't just obeying as human wanted it only when told to buy groceries? Well we then train this AGI on other tasks. When we are sufficiently confident that this AGI obeys as human wanted it in other tasks, we deploy it.

But hold on, in the real world the AGI can access the real uncurated internet, learn about hacking and the real stock market. Note that this AGI is never trained on hacking training in the simulation, as simulating the internet is a bit too much.

Now, he is asked by his owner to buy a gold bar for as cheap as possible. Hacking an online shop to get a gold bar is a perfectly valid strategy! Because he is never trained in this scenario before, thus the moral restriction is not specified.

I think your argument hinges on the fact that morality will get generalized outside of the training environment. Which might or might not be true. This is becoming even more complex with the fact that AGI might found solutions which is not just excluded in training simulation, but also have never been considered by humanity as a whole. New technology for example.

1

WarImportant9685 t1_j16uzo7 wrote

I think you generalized current AI to AGI. The most useful trait of AGI, but also the most problematic is that it can self-learn.

So that the training environment can be much smaller than the real world. But then, if the training environment is so small, how can we be sure that human morals/obedience is generalized to the real world?

What kind of reward function/training process would elicit generalization of the expected behaviour?

I would like to hear your thoughts about this.

1

WarImportant9685 t1_j0ngwf5 wrote

I understand your point. Although we are not on the same page, I believe we are on the same chapter.

I think my main disagreement is that to recognize undesirable 'thoughts' in AI is not such an easy problem. As from my previous comments, one of the holy grail of AI interpretation study is detecting a lying AI which mean we are talking about the same thing! But you are more optimistic than I do, which is fine.

I also understand that we might be able design the AI to use less black-boxy structure to aid AI interpretation. But again I'm not too optimistic about this. I just have no idea how it can be achieved. As at a glance it seems like they are on different abstraction levels. Like if we are just designing the building blocks. How can we dictate how it is going to be used.

Like how are you supposed to design lego blocks, so that it cannot be used to create dragons.

Then again, maybe I'm just too doomer, as alignment problem is unsolved, AGI haven't been solved too. So I agree with you, we'll have to see how it goes.

1

WarImportant9685 t1_j0muh8q wrote

I do hope I share your optimism. But from the research I read, it seems that even the control problem seems to be a hard problem for us right now. As a fellow researcher what makes you personally feel optimistic that it'll be easy to solve?

I'll try to take a shot why I think the solution you said, is likely to be moot.

Direct modification of memory -> This is an advantage yes. But it's useless if we don't understand the AI in the way that we want. For the holy grail ideally we can understand if the AI is lying by looking at the neural weights. Or maybe searching with 100% certainty if the AI have mesa-optimizer for its subroutine. But our current AI interpretability research is still so far away from that.

Seamless sandboxing -> I'm not sure what you mean by this. But if I was to take a shot, I'll interpret this as true simulation of the real world. Which is impossible! My reasoning is that, the real world doesn't only contain garden, lake, and atom interactions. But also tons of human doing what the fuck they usually did. The economics and so on and on. What we can get is only 'close enough' simulation. But how do we define close enough? No one knows how to define this rigorously

Soft influence -> Not sure what you mean by this

Hard behavior modification -> I'll interpret this as hard rules for the AI to follow? Not gonna work. There is a reason why we are moving on from expert systems to AI. And we want to control AI with expert systems?

And anyway, I do want to hear your reply as a fellow researcher. Hopefully I don't come across as rude

1

WarImportant9685 t1_j0mis5x wrote

yeah, IF we succeed in AI alignment to singular entity whether it is corporation/singular human. The question become the age old question of humankind that is, greed or altruism?

What will the entity that gain the power first do, I'm more inclined to think that we are just incapable of knowing the answer yet. As it is too situational to whom the power is attained first.

Unaligned AGI is a whole different beast tho.

1

WarImportant9685 t1_j0m2a4i wrote

no he's correct, inequality has been rising insanely because of technology. Of course the socioeconomic thing affects that too. But tech definitely amplified it. As it makes it easier for the rich to get richer. Easiest example are billionares.

But you're also correct that the lower and middle class standard of living has been uplifted by technology.

I think the question is, if when AGI arrived, what will happen, there are several views, which two of them corresponding to OP and you:

  1. the inequality become insane insane insane. As the lower and middle class replaced by AGI. They are gone/die by starvation/last struggle
  2. Maybe it's the contrary, as it's extremely easy to get resources. The lower and middle class standard of living is lifted to be very good. Maybe the abolishment of the class altogether.

IMHO, we don't know which one will happen. Thus dubbed the singularity. As possibly for the first time in history we are going to create a thinking machine that are much smarter than human. Whether it will be the god or the devil or just the staff of god or the trident of the devil. No body knows. But the alignment theory does teach us the possibility of the devil is real and logical.

−2

WarImportant9685 t1_j0m1flg wrote

yeah this concerns me. Blatant stealing of art from known artists. With seemingly no public backlash from the tech community seems distasteful. Even though we (the tech community) are the one that understood, that the training data must have been webcrawled from the artists with no permisson. It seems kinda trashy that we don't care about other community as long as it doesn't touch our territory.

I've always identified with tech people. AI makes me think twice.

−4

WarImportant9685 t1_iwmeray wrote

Reply to comment by RavenWolf1 in Ai art is a mixed bag by Nintell

Easy to say, but people in general likes shitting on antiwork. With some argument that they earned their money, how come other people can get basic income freely? Of course, then there's the argument about communism.

2