Viewing a single comment thread. View all comments

Zer0pede t1_j7vtopj wrote

Not if the “safeguards” are structured like a value system. I like the approach in Stuart Russel’s “Human Compatible,” which is that we start now making AI have the same “goals”* as humans (including checking with humans to confirm).

*I put “goals” in quotes because it makes AI sound conscious, but literally no AI researcher is working on consciousness so we’re really just talking about a system that “checks in” with humans to make sure it doesn’t achieve a minor human-assigned goal at the expense of more important, abstract human values. (eg., Paperclip Multiplier or Facebook algorithm.)

2