Viewing a single comment thread. View all comments

Heinrick_Veston t1_jefhmai wrote

Could we not just hard code these models to constantly ask if they're behaving properly and doing what we want?

Perhaps we could have a democratic system we use to respond to those queries, to make sure they best represent us all.

2

DaggerShowRabs t1_jefm4ex wrote

If the system needs approval before it takes any actions at all, the system is going to be extremely slow and limited.

2

Heinrick_Veston t1_jefmvuu wrote

I don't mean that it would ask before every action, more so that it'd regularly ask if it was acting in the right way.

1

DaggerShowRabs t1_jefnl06 wrote

Ah, I get what you mean. I still don't think that necessarily solves the problem. It could be possible for a hypothetical artificial superintelligence to take actions that seem harmless to us, but because it is better at planning and prediction than us, the system knows the action or series of actions will lead to humanity's demise. But since it appears harmless to us, when it asks, we say, "Yes, you are acting in the correct way".

3