Viewing a single comment thread. View all comments

NotARedditUser3 t1_j90j0er wrote

If you spend some time looking up how microsoft's gpt integrated chat / ai works, it does this. Lookup the thread of tweets for the hacker that exposed its internal codename 'Syndey'; it scrapes his twitter profile, realizes he exposed its secrets in prior convo's after social engineering hacking it with a few conversations, and then turns hostile to him.

1

a1_jakesauce_ t1_j90k4h6 wrote

1

blablanonymous t1_j917xm2 wrote

Is that real? I don’t know why I feel like it could be totally fake

2

currentscurrents t1_j96vbfj wrote

Microsoft has confirmed the rules are real:

>We asked Microsoft about Sydney and these rules, and the company was happy to explain their origins and confirmed that the secret rules are genuine.

The rest, who knows. I never got access before they fixed it. But there are many screenshots from different people of it acting quite unhinged.

2

blablanonymous t1_j96xu8w wrote

Thanks for the link!

I mean I guess there was nothing too surprising about the rules, given how these systems work (essentially trying to predict the end of a user input text). But the rest, seems so ridiculously dramatic that I wouldn’t be shocked if he specifically prompted it to be that dramatic and hid that part. I’m probably being paranoid, since at least the rules part is true, but it seems like the perfect conversation to elicit every single fear people have about AI.

1

NotARedditUser3 t1_j9225b5 wrote

I'll reply back with what I was referring to later, it was a different thing

0