home bbs files messages ]

Forums before death by AOL, social media and spammers... "We can't have nice things"

   alt.cyberpunk.tech      Cyberpunks LOVE making shit complicated      1,115 messages   

[   << oldest   |   < older   |   list   |   newer >   |   newest >>   ]

   Message 290 of 1,115   
   memstoner to flashheart   
   Re: Chat GPT has now fully automated red   
   13 Oct 25 11:23:42   
   
   From: memstoner@420blaze.it   
      
   On Sun, 12 Oct 2025 17:31:02 +0300   
   flashheart  wrote:   
      
   > > As AI bots like ChatGPT become inextricably tangled with people’s   
   > > private and public lives, it’s causing unpredictable new crises.   
   > >    
   > > One of these collision points is in romantic relationships, where   
   > > an uncanny dynamic is unfolding across the world: one person in a   
   > > couple becomes fixated on ChatGPT or another bot — for some   
   > > combination of therapy, relationship advice, or spiritual wisdom —   
   > > and ends up tearing the partnership down as the AI makes more and   
   > > more radical interpersonal suggestions.     
   >    
   > > In one chaotic recording we obtained, two married women are inside   
   > > a moving car, their two young children sitting in the backseat.   
   > >    
   > > The tension in the vehicle is palpable. The marriage has been on   
   > > the rocks for months, and the wife in the passenger seat, who   
   > > recently requested an official separation, has been asking her   
   > > spouse not to fight with her in front of their kids. But as the   
   > > family speeds down the roadway, the spouse in the driver’s seat   
   > > pulls out a smartphone and starts quizzing ChatGPT’s Voice Mode   
   > > about their relationship problems, feeding the chatbot leading   
   > > prompts that result in the AI browbeating her wife in front of   
   > > their preschool-aged children.   
   > >    
   > > After funneling her complaints into ChatGPT, the driver asks the   
   > > bot to analyze the prompts as if “a million therapists” were going   
   > > to “read and weigh in.”   
   > >    
   > > “The responses you’ve described would likely be considered unfair   
   > > and emotionally harmful by the majority of marriage therapists,”   
   > > the chatbot responds at a loud volume, while mirroring back the   
   > > same language used in the prompt with flowery therapy-speak. It   
   > > offers no pushback, nor does it attempt to reframe the driver’s   
   > > perspective. At one point, the chatbot accuses the wife in the   
   > > passenger seat of engaging in “avoidance through boundaries” by   
   > > requesting that they not fight in front of their kids — while those   
   > > very children sit in the vehicle, just feet away.   
   > >    
   > > It goes on and on, with ChatGPT monologuing while the wife it’s   
   > > being wielded against occasionally tries to cut in over its robotic   
   > > lecture. The spouse prompting the bot, meanwhile, mutters approving   
   > > commentary: “that’s right,” “mm-hmm,” “see?”   
   > >    
   > > “Please keep your eyes on the road,” the wife being lectured by the   
   > > AI pleads at one point.   
   > >    
   > > This was a regular occurrence, she told us, in which her spouse   
   > > would pull out ChatGPT and prompt it to agree with her in   
   > > long-winded diatribes.   
   > >    
   > > “We were arguing a lot… we would be up all night, and I would   
   > > assert a boundary, or say, like, ‘I don’t want to have this   
   > > discussion in front of the kids,’ or ‘I need to go to bed,’” she   
   > > recounted, “and [my ex] would immediately turn on ChatGPT and start   
   > > talking to it, and be like, ‘can you believe what she’s doing?'”   
   > >    
   > > Her ex would carry out these conversations with ChatGPT on speaker   
   > > phone, she added — within earshot, pointedly, so she could hear   
   > > everything.   
   > >    
   > > “[My ex] would have it on speaker phone, and then have it speak not   
   > > to me, but it would be in the same room,” she recalled. “And of   
   > > course, ChatGPT was this confirmative voice, being like, ‘you’re so   
   > > right.'”   
   > >    
   > > Today, the former couple, together nearly 15 years, is in the midst   
   > > of a contentious divorce and custody battle.     
   >    
   > > Even Geoffrey Hinton, a Nobel Prize-winning computer scientist   
   > > known as a “Godfather of AI” — a technology that likely wouldn’t   
   > > exist in its current form without his contributions — recently   
   > > conceded that his girlfriend had broken up with him using ChatGPT.   
   > >    
   > > “She got ChatGPT to tell me what a rat I was… she got the chatbot   
   > > to explain how awful my behavior was and gave it to me,” Hinton   
   > > told The Financial Times. “I didn’t think I had been a rat, so it   
   > > didn’t make me feel too bad.”     
   >    
   > and many more examples of LLM induced retardation @    
   > https://futurism.com/chatgpt-marriages-divorces   
   >    
   > We are currently careening towards a future where people will become    
   > increasingly solitary because some autocomplete trained on billions   
   > of reddit comments is telling them to terminate their relationships.   
   > This is clearly happening on accident, I don't think Sam Altman    
   > personally wants people breaking up because of his chatbot.   
   > However, now that Pandora's box is opened, do you think that   
   > Generative AI companies will begin manipulating people towards   
   > atomization? Love and the fulfillment derived from it is one of the   
   > few things that cannot be bought and sold in this increasingly   
   > financialized present. There are powerful interests against people   
   > deriving enjoyment out of something that isn't a product.   
      
   It's just plain stupid  to ask an AI (which can state the fact that   
   they have no emotions or perspective the way we have) to ask about   
   problems. Are people buying into this?    
      
   --- SoupGate-Win32 v1.05   
    * Origin: you cannot sedate... all the things you hate (1:229/2)   

[   << oldest   |   < older   |   list   |   newer >   |   newest >>   ]


(c) 1994,  bbs@darkrealms.ca