home bbs files messages ]

Forums before death by AOL, social media and spammers... "We can't have nice things"

   alt.cyberpunk.tech      Cyberpunks LOVE making shit complicated      1,115 messages   

[   << oldest   |   < older   |   list   |   newer >   |   newest >>   ]

   Message 246 of 1,115   
   MasterKarsten to flashheart   
   Re: Chat GPT has now fully automated red   
   12 Oct 25 16:33:53   
   
   From: soulkarsten@gmail.com   
      
   TheSun, 12 Oct 2025 17:31:02 +0300, flashheart wrote:   
      
   >> As AI bots like ChatGPT become inextricably tangled with people’s   
   >> private and public lives, it’s causing unpredictable new crises.   
   >>   
   >> One of these collision points is in romantic relationships, where an   
   >> uncanny dynamic is unfolding across the world: one person in a couple   
   >> becomes fixated on ChatGPT or another bot — for some combination of   
   >> therapy, relationship advice, or spiritual wisdom — and ends up tearing   
   >> the partnership down as the AI makes more and more radical   
   >> interpersonal suggestions.   
   >   
   >> In one chaotic recording we obtained, two married women are inside a   
   >> moving car, their two young children sitting in the backseat.   
   >>   
   >> The tension in the vehicle is palpable. The marriage has been on the   
   >> rocks for months, and the wife in the passenger seat, who recently   
   >> requested an official separation, has been asking her spouse not to   
   >> fight with her in front of their kids. But as the family speeds down   
   >> the roadway, the spouse in the driver’s seat pulls out a smartphone and   
   >> starts quizzing ChatGPT’s Voice Mode about their relationship problems,   
   >> feeding the chatbot leading prompts that result in the AI browbeating   
   >> her wife in front of their preschool-aged children.   
   >>   
   >> After funneling her complaints into ChatGPT, the driver asks the bot to   
   >> analyze the prompts as if “a million therapists” were going to “read   
   >> and weigh in.”   
   >>   
   >> “The responses you’ve described would likely be considered unfair and   
   >> emotionally harmful by the majority of marriage therapists,” the   
   >> chatbot responds at a loud volume, while mirroring back the same   
   >> language used in the prompt with flowery therapy-speak. It offers no   
   >> pushback, nor does it attempt to reframe the driver’s perspective. At   
   >> one point, the chatbot accuses the wife in the passenger seat of   
   >> engaging in “avoidance through boundaries” by requesting that they not   
   >> fight in front of their kids — while those very children sit in the   
   >> vehicle, just feet away.   
   >>   
   >> It goes on and on, with ChatGPT monologuing while the wife it’s being   
   >> wielded against occasionally tries to cut in over its robotic lecture.   
   >> The spouse prompting the bot, meanwhile, mutters approving commentary:   
   >> “that’s right,” “mm-hmm,” “see?”   
   >>   
   >> “Please keep your eyes on the road,” the wife being lectured by the AI   
   >> pleads at one point.   
   >>   
   >> This was a regular occurrence, she told us, in which her spouse would   
   >> pull out ChatGPT and prompt it to agree with her in long-winded   
   >> diatribes.   
   >>   
   >> “We were arguing a lot… we would be up all night, and I would assert a   
   >> boundary, or say, like, ‘I don’t want to have this discussion in front   
   >> of the kids,’ or ‘I need to go to bed,’” she recounted, “and [my   
   ex]   
   >> would immediately turn on ChatGPT and start talking to it, and be like,   
   >> ‘can you believe what she’s doing?'”   
   >>   
   >> Her ex would carry out these conversations with ChatGPT on speaker   
   >> phone, she added — within earshot, pointedly, so she could hear   
   >> everything.   
   >>   
   >> “[My ex] would have it on speaker phone, and then have it speak not to   
   >> me, but it would be in the same room,” she recalled. “And of course,   
   >> ChatGPT was this confirmative voice, being like, ‘you’re so right.'”   
   >>   
   >> Today, the former couple, together nearly 15 years, is in the midst of   
   >> a contentious divorce and custody battle.   
   >   
   >> Even Geoffrey Hinton, a Nobel Prize-winning computer scientist known as   
   >> a “Godfather of AI” — a technology that likely wouldn’t exist in its   
   >> current form without his contributions — recently conceded that his   
   >> girlfriend had broken up with him using ChatGPT.   
   >>   
   >> “She got ChatGPT to tell me what a rat I was… she got the chatbot to   
   >> explain how awful my behavior was and gave it to me,” Hinton told The   
   >> Financial Times. “I didn’t think I had been a rat, so it didn’t make   
   me   
   >> feel too bad.”   
   >   
   > and many more examples of LLM induced retardation @   
   > https://futurism.com/chatgpt-marriages-divorces   
   >   
   > We are currently careening towards a future where people will become   
   > increasingly solitary because some autocomplete trained on billions of   
   > reddit comments is telling them to terminate their relationships.   
   > This is clearly happening on accident, I don't think Sam Altman   
   > personally wants people breaking up because of his chatbot.   
   > However, now that Pandora's box is opened, do you think that Generative   
   > AI companies will begin manipulating people towards atomization?   
   > Love and the fulfillment derived from it is one of the few things that   
   > cannot be bought and sold in this increasingly financialized present.   
   > There are powerful interests against people deriving enjoyment out of   
   > something that isn't a product.   
      
      
   It's the people not the LLM... Why one should ask certain stuff to AI is   
   beyond me   
      
      
   --   
   ---/g/ on Usenet? Madness!---   
      
   --- SoupGate-Win32 v1.05   
    * Origin: you cannot sedate... all the things you hate (1:229/2)   

[   << oldest   |   < older   |   list   |   newer >   |   newest >>   ]


(c) 1994,  bbs@darkrealms.ca