- cross-posted to:
- technology@lemmit.online
- cross-posted to:
- technology@lemmit.online
AI girlfriend bots are already flooding OpenAI’s GPT store::OpenAI’s store rules are already being broken, illustrating that regulating GPTs could be hard to control
AI girlfriend bots are already flooding OpenAI’s GPT store::OpenAI’s store rules are already being broken, illustrating that regulating GPTs could be hard to control
[Yawn]
I’m all for a bit of Ai panic, but this is the worst kind of desperate journalism.
The facts as reported:
The articles conclusion: Ai regulation is doomed to fail and the machines will wipe out humanity.
Well, as we all know, AI girlfriend is the first step to AI Hitler.
Stupid sexy hitler UwU.
deleted by creator
A very solid point :-)
If we get wiped out by AI girlfriends we deserve it. If the reason why a person never reproduced is solely because they had a chatbot they really should not reproduce.
I was trying to dream up the justification for this rule that wasn’t about mitigating the ick-factor and fell short… I guess if the machines learn how to beguile us by forming relationships then they could be used to manipulate people honeypot style?
Honestly the only point I set out to make was that people were probably working on virtual girlfriends for weeks (months?) before they were banned. They had probably been submitted to the store already and the article was trying to drum up panic.
Sure which you know we already can do. Honeypots are a thing and a thing so old the Bible mentions them. Delilah anyone? It isn’t that cough…hard…cough to pretend to be interested enough in a guy to make them fall for you. Sure if the tech keeps growing, which it will, you can imagine more and more complex cons. Stuff that could even have webcam chats with the marks.
I suggest we treat this the same way we currently treat humans doing this. We warn users, block accounts that do this, and criminally prosecute.
Its a hard question to answer, there is a good reason but its sevral pargraphs long and i likely have gaps in knolage and in some places misguided. The reduced idea: being emotionally open (no emotional guarding or sandboxing/RPing) with a creature that lacks many traits required to take on that responsability. the model is being pretrained to perform jestures that make us happy, having no internal state to ask itself if it would enjoy garlic bread given its experience with garlic. its an advanced tape recorder, being pre-populated with an answer. Or it lies and picks somthing because saying idk is the wrong response. As apposed to a creature that has some kind of consistant external world and a memory system. firehosing it with data, means less room for artistic intent.
If your sandboxing/Roleplaying, theres no problem.
Interesting idea. We could effectively practice eugenics in a way that won’t make people so mad. They’ll have to contend with ideas like free will and personal responsibility before they can go after our program.
Let’s make a list of all the “asocials” we want removed from the gene pool and we can get started.
Not that I’m really interested in one but what’s actually wrong with making an AI gf app?
It encourages the dehumanization of women and gives men even more unrealistic expections about relationships and sex. But if they take themselves out of the gene pool this way then it could end up being a win.
As if dehumanization of men wasn’t just as bad.
You know, I saw a pic posted somewhere recently saying something about not liking bodybuilders and unrealistically cool guys, those she likes are absolutely normal and casual, like guys on the picture.
And guys on the picture are Hollywood actors, LOL, in very good form, with no signs of sleep deprivation and tiredness, with a selling smile and the photos are likely edited on top of that.
And the totally realistic and normal expectation of many women towards men is that if a woman has a moment of weakness and pain, then it’s her personality to be proud of, and if a man has that, then he should accept being dumped for that moment alone as a man.
I actually think it absolutely mirrors the dehumanization of women. All the same things.
You are losing sight of the discussion to frame it as a “men vs women” thing. This will also feed into the dehumanization of men because it will also generated “ideal” impossible men.
N-nah. But if we get back to the root of this discussion - I’ve read lots of fanfiction in my life. Mostly written by girls for girls. Taboons of imagined idealized men right there.
And about imagined idealized women - men write fanfiction (and other fiction) too.
So I just don’t see how such bots are bad, except they are not real.
You know, i’ve noticed over my 40+ years that the vast majority of men are unnattractive. Men like to rate women on a scale but i just do a yes or no and 97% are no. But they still get girlfriends, get married and have kids. Ignore the women who care about looks because they seem to be a tiny minority.
I didn’t have to do that anyway, my problems in this area result mostly from my own mistakes, but one can’t just abruptly stop making them.
Though I think I actually get something right, after the dust settles I still rather like (as people) everybody for whom I felt something.
They could train it however they want, it wouldn’t have to be dehumanizing (admittedly probably wouldn’t be as successful). Hell, maybe they could disguise a therapy AI as a gf AI and trick them into getting their shit together.
Side question, how do you feel about romance movies/novels that give unrealistic expectations of men? Should those be banned as well?
I never said it should be banned, just that I don’t like it.
People will train it in all kinds of ways. Lets take sex out of tbe equation and say a nazi trains a home automation/personal assistant ai as a house removed bot. Still cool?
I’m not sure there’s anything wrong with it, that’s what the article reported on as though it were some sort of harbinger of doom… Felt like my smarmy retorts would be slightly less punchy if I had opened up a side discussion regarding appropriate uses for AI. I suppose part of my motivation was that it seemed incredibly innocent relatively speaking.
Open AI claims to be in this to save humanity from Skynet, this seems like a fairly pathetic attempt keep their store from filling up with “disreputable” content before… what exactly I don’t know. The killer app for AI that would be magically devoid of controversy?
Are people really wargaming this? Planning on making anti-skynets to defend humanity from skynets? I can’t decide if that is a massive waste of time or a vital use of it.
I’m not an authority on the subject, but that was my understanding from the reporting surrounding Open AI’s recent kerfuffle. That their complex management structure was part of some elaborate strategy to promote the development of ethical AI.
Sounded a bit sus to me, but clearly smarter folks think it’s a good way to spend money.
Your comment gave me an idea. These alarmist articles are so common that I bet writing them could be automated. We can get bots to write articles about the dangers of bots. I asked chatgpt to write one from the perspective of Southern Baptist
Not bad for a first pass.
LOL
That’s kind of fascinating, because I think it authentically feels like it might be the perspective behind some fire-and-brimstone speech on the subject. I was kind of hopping for the sermon personally, but this makes you feel like southern baptist preachers could be people too ;-)