Forums before death by AOL, social media and spammers... "We can't have nice things"
|    alt.cyberpunk.tech    |    Cyberpunks LOVE making shit complicated    |    1,115 messages    |
[   << oldest   |   < older   |   list   |   newer >   |   newest >>   ]
|    Message 246 of 1,115    |
|    MasterKarsten to flashheart    |
|    Re: Chat GPT has now fully automated red    |
|    12 Oct 25 16:33:53    |
      From: soulkarsten@gmail.com              TheSun, 12 Oct 2025 17:31:02 +0300, flashheart wrote:              >> As AI bots like ChatGPT become inextricably tangled with people’s       >> private and public lives, it’s causing unpredictable new crises.       >>       >> One of these collision points is in romantic relationships, where an       >> uncanny dynamic is unfolding across the world: one person in a couple       >> becomes fixated on ChatGPT or another bot — for some combination of       >> therapy, relationship advice, or spiritual wisdom — and ends up tearing       >> the partnership down as the AI makes more and more radical       >> interpersonal suggestions.       >       >> In one chaotic recording we obtained, two married women are inside a       >> moving car, their two young children sitting in the backseat.       >>       >> The tension in the vehicle is palpable. The marriage has been on the       >> rocks for months, and the wife in the passenger seat, who recently       >> requested an official separation, has been asking her spouse not to       >> fight with her in front of their kids. But as the family speeds down       >> the roadway, the spouse in the driver’s seat pulls out a smartphone and       >> starts quizzing ChatGPT’s Voice Mode about their relationship problems,       >> feeding the chatbot leading prompts that result in the AI browbeating       >> her wife in front of their preschool-aged children.       >>       >> After funneling her complaints into ChatGPT, the driver asks the bot to       >> analyze the prompts as if “a million therapists” were going to “read       >> and weigh in.”       >>       >> “The responses you’ve described would likely be considered unfair and       >> emotionally harmful by the majority of marriage therapists,” the       >> chatbot responds at a loud volume, while mirroring back the same       >> language used in the prompt with flowery therapy-speak. It offers no       >> pushback, nor does it attempt to reframe the driver’s perspective. At       >> one point, the chatbot accuses the wife in the passenger seat of       >> engaging in “avoidance through boundaries” by requesting that they not       >> fight in front of their kids — while those very children sit in the       >> vehicle, just feet away.       >>       >> It goes on and on, with ChatGPT monologuing while the wife it’s being       >> wielded against occasionally tries to cut in over its robotic lecture.       >> The spouse prompting the bot, meanwhile, mutters approving commentary:       >> “that’s right,” “mm-hmm,” “see?”       >>       >> “Please keep your eyes on the road,” the wife being lectured by the AI       >> pleads at one point.       >>       >> This was a regular occurrence, she told us, in which her spouse would       >> pull out ChatGPT and prompt it to agree with her in long-winded       >> diatribes.       >>       >> “We were arguing a lot… we would be up all night, and I would assert a       >> boundary, or say, like, ‘I don’t want to have this discussion in front       >> of the kids,’ or ‘I need to go to bed,’” she recounted, “and [my       ex]       >> would immediately turn on ChatGPT and start talking to it, and be like,       >> ‘can you believe what she’s doing?'”       >>       >> Her ex would carry out these conversations with ChatGPT on speaker       >> phone, she added — within earshot, pointedly, so she could hear       >> everything.       >>       >> “[My ex] would have it on speaker phone, and then have it speak not to       >> me, but it would be in the same room,” she recalled. “And of course,       >> ChatGPT was this confirmative voice, being like, ‘you’re so right.'”       >>       >> Today, the former couple, together nearly 15 years, is in the midst of       >> a contentious divorce and custody battle.       >       >> Even Geoffrey Hinton, a Nobel Prize-winning computer scientist known as       >> a “Godfather of AI” — a technology that likely wouldn’t exist in its       >> current form without his contributions — recently conceded that his       >> girlfriend had broken up with him using ChatGPT.       >>       >> “She got ChatGPT to tell me what a rat I was… she got the chatbot to       >> explain how awful my behavior was and gave it to me,” Hinton told The       >> Financial Times. “I didn’t think I had been a rat, so it didn’t make       me       >> feel too bad.”       >       > and many more examples of LLM induced retardation @       > https://futurism.com/chatgpt-marriages-divorces       >       > We are currently careening towards a future where people will become       > increasingly solitary because some autocomplete trained on billions of       > reddit comments is telling them to terminate their relationships.       > This is clearly happening on accident, I don't think Sam Altman       > personally wants people breaking up because of his chatbot.       > However, now that Pandora's box is opened, do you think that Generative       > AI companies will begin manipulating people towards atomization?       > Love and the fulfillment derived from it is one of the few things that       > cannot be bought and sold in this increasingly financialized present.       > There are powerful interests against people deriving enjoyment out of       > something that isn't a product.                     It's the people not the LLM... Why one should ask certain stuff to AI is       beyond me                     --       ---/g/ on Usenet? Madness!---              --- SoupGate-Win32 v1.05        * Origin: you cannot sedate... all the things you hate (1:229/2)    |
[   << oldest   |   < older   |   list   |   newer >   |   newest >>   ]
(c) 1994, bbs@darkrealms.ca