Submitted by BronzeArcher t3_1150kh0 in MachineLearning
NotARedditUser3 t1_j8z5dy8 wrote
Imagine someone writes one that's explicitly aimed around manipulating your thoughts and actions.
An AI could likely come up with some insane tactics for this. Could feed off of your twitter page, find an online resume of you or scrape other social media or in microsoft's case or google's, potentially scrape your emails you have with them, profile you in an instant, and then come up with a tailor made advertisement or argument that it knows would land on you.
Scary thought.
pyepyepie t1_j93iif6 wrote
Honestly, much simpler algorithms already do it to some extent (recommendation systems), the biggest difference is that it has to suggest you a post someone else wrote instead of writing it by itself. Great take :)
Philiatrist t1_j8zwxbg wrote
How would the AI know it’s profiling you and not the other AI you’ve set up to do all of those things for you?
BronzeArcher OP t1_j8z7vfi wrote
Yeah that’s pretty frightening.
[deleted] t1_j8zbtif wrote
[removed]
[deleted] t1_j917vlr wrote
[deleted]
currentscurrents t1_j8zwnht wrote
It depends on whether it's exploiting my psychology to sell me something I don't need, or if it's gathering information to find something that may actually be useful for me. I suspect the latter is a more useful strategy in the long run because people tend to adjust to counter psychological exploits.
If I'm shown an advertisement for something I actually want... that doesn't sound bad? I certainly don't like ads for irrelevant things like penis enlargement.
sweetchocolotepie t1_j91vuca wrote
there is no "useful vs unuseful", you either want it or do not want it. the usefulness is something you define which is subset of the things you want. however the model will just suggest you stuff that may or may not be practical to you, but you want it. you may find them pseudo-useful or useful at the moment or....
case is, it will sell
a1_jakesauce_ t1_j90iv0q wrote
This describes a LLM + reinforcement learning hybrid that has been trained to navigate webpages for arbitrary tasks. I’m not sure how far away this is, or if it already exists. Someone below mentioned an action transformer which may be related
NotARedditUser3 t1_j90j0er wrote
If you spend some time looking up how microsoft's gpt integrated chat / ai works, it does this. Lookup the thread of tweets for the hacker that exposed its internal codename 'Syndey'; it scrapes his twitter profile, realizes he exposed its secrets in prior convo's after social engineering hacking it with a few conversations, and then turns hostile to him.
a1_jakesauce_ t1_j90k4h6 wrote
All I found was this https://twitter.com/kliu128/status/1623472922374574080?s=21
blablanonymous t1_j917xm2 wrote
Is that real? I don’t know why I feel like it could be totally fake
currentscurrents t1_j96vbfj wrote
Microsoft has confirmed the rules are real:
>We asked Microsoft about Sydney and these rules, and the company was happy to explain their origins and confirmed that the secret rules are genuine.
The rest, who knows. I never got access before they fixed it. But there are many screenshots from different people of it acting quite unhinged.
blablanonymous t1_j96xu8w wrote
Thanks for the link!
I mean I guess there was nothing too surprising about the rules, given how these systems work (essentially trying to predict the end of a user input text). But the rest, seems so ridiculously dramatic that I wouldn’t be shocked if he specifically prompted it to be that dramatic and hid that part. I’m probably being paranoid, since at least the rules part is true, but it seems like the perfect conversation to elicit every single fear people have about AI.
[deleted] t1_j96v1wv wrote
[deleted]
NotARedditUser3 t1_j9225b5 wrote
I'll reply back with what I was referring to later, it was a different thing
Viewing a single comment thread. View all comments