Viewing a single comment thread. View all comments

Smack-works OP t1_itn7xuq wrote

You make way too many assumptions and inferences at every single turn of your answers. You don't question those assumptions and inferences. And you don't make those assumptions and inferences clear so that I can comfortably agree/disagree with them. You make no effort to check if we are on the same page or not.

> When you represent that you can provide knowledge from a set of statements, but the dataset they are meant to represent is an infinite one, the first thing you are establishing is that the finite data that you have won't really be representative - so you won't be able to make behavior guarantees.

As I understand your reasoning chain: "X statements are an infinite set = AI needs to know the entire set to be aligned = we need infinite memory for this".

Do you realize that this reasoning chain contains at least 2 assumptions which can be fully or partially wrong?

> In order for this set of information to not be infinite, the set requires properties that are true for all the statements in the set ie that it is possible to measure and validate if a statement should be inside or outside the set. Having a validity check means that the second value statement that you try to add to the set cannot be arbitrary - because an arbitrary statement may well be contradictory to the first statement.

You are potentially confusing 4 different things:

  • Set of all X statements. Set of the true X statements. Set of the statements the AI needs to know. Checking if a statement is X statement or not. Checking if X statement is true or not.

Not saying you are actually confused. But what you write doesn't differentiate between those 4 things. So answering to what you wrote is extremely frustrating.

> 'You don't need to solve ethics in order to learn values.' How do you learn values then? If you don't know, then you are also saying you don't know how to learn X-statements.

People learn values without solving ethics.

1

gahblahblah t1_itoqqsi wrote

Lots of communication involves making reasonable assumptions, so that a person doesn't need to spell out every detail. My presumptions are only a problem if they are wrong.

'People learn values without solving ethics'.

Your non-answer answer to my question leads me to conclude that I am wasting time trying to ask you further questions, so we can let it all go.

1

Smack-works OP t1_itou8ft wrote

Continuing to make presumptions (when you see that the previous ones are not clear) may be a problem too.

But I think your assumptions are wrong:

> The first trouble with your X statements, is they seem like an infinite set. The examples you give for your X statements in point 3 don't seem to come from a finite list of statements that you could just hand to a system. Rather they appear to be rational that you'd explain after encountering a specific situation.

  • AI doesn't need to know an infinite set of X statements.
  • You don't need to give all of the statements to the system beforehand. It can learn them.
  • It's OK if some statements are deduced after encountering a specific situation.

X statements are not supposed to encode the absolute ethics we put into the system beforehand.

> 'People learn values without solving ethics'.

> Your non-answer answer to my question leads me to conclude that I am wasting time trying to ask you further questions, so we can let it all go.

You assume I'm supposed to answer to everything you write? I wanted you to admit at least some common truth ("we don't need to solve ethics") before dealing with more assumptions and inferences.

1