Forums before death by AOL, social media and spammers... "We can't have nice things"
|    alt.cyberpunk.tech    |    Cyberpunks LOVE making shit complicated    |    1,115 messages    |
[   << oldest   |   < older   |   list   |   newer >   |   newest >>   ]
|    Message 256 of 1,115    |
|    MasterKarsten to flashheart    |
|    Re: Chat GPT has now fully automated red    |
|    12 Oct 25 16:56:18    |
      From: soulkarsten@gmail.com              TheSun, 12 Oct 2025 19:43:56 +0300, flashheart wrote:              > On 10/12/25 19:33, MasterKarsten wrote:       >> TheSun, 12 Oct 2025 17:31:02 +0300, flashheart wrote:       >>       >>>> As AI bots like ChatGPT become inextricably tangled with people’s       >>>> private and public lives, it’s causing unpredictable new crises.       >>>>       >>>> One of these collision points is in romantic relationships, where an       >>>> uncanny dynamic is unfolding across the world: one person in a couple       >>>> becomes fixated on ChatGPT or another bot — for some combination of       >>>> therapy, relationship advice, or spiritual wisdom — and ends up       >>>> tearing the partnership down as the AI makes more and more radical       >>>> interpersonal suggestions.       >>>       >>>> In one chaotic recording we obtained, two married women are inside a       >>>> moving car, their two young children sitting in the backseat.       >>>>       >>>> The tension in the vehicle is palpable. The marriage has been on the       >>>> rocks for months, and the wife in the passenger seat, who recently       >>>> requested an official separation, has been asking her spouse not to       >>>> fight with her in front of their kids. But as the family speeds down       >>>> the roadway, the spouse in the driver’s seat pulls out a smartphone       >>>> and starts quizzing ChatGPT’s Voice Mode about their relationship       >>>> problems, feeding the chatbot leading prompts that result in the AI       >>>> browbeating her wife in front of their preschool-aged children.       >>>>       >>>> After funneling her complaints into ChatGPT, the driver asks the bot       >>>> to analyze the prompts as if “a million therapists” were going to       >>>> “read and weigh in.”       >>>>       >>>> “The responses you’ve described would likely be considered unfair and       >>>> emotionally harmful by the majority of marriage therapists,” the       >>>> chatbot responds at a loud volume, while mirroring back the same       >>>> language used in the prompt with flowery therapy-speak. It offers no       >>>> pushback, nor does it attempt to reframe the driver’s perspective. At       >>>> one point, the chatbot accuses the wife in the passenger seat of       >>>> engaging in “avoidance through boundaries” by requesting that they       >>>> not fight in front of their kids — while those very children sit in       >>>> the vehicle, just feet away.       >>>>       >>>> It goes on and on, with ChatGPT monologuing while the wife it’s being       >>>> wielded against occasionally tries to cut in over its robotic       >>>> lecture.       >>>> The spouse prompting the bot, meanwhile, mutters approving       >>>> commentary: “that’s right,” “mm-hmm,” “see?”       >>>>       >>>> “Please keep your eyes on the road,” the wife being lectured by the       >>>> AI pleads at one point.       >>>>       >>>> This was a regular occurrence, she told us, in which her spouse would       >>>> pull out ChatGPT and prompt it to agree with her in long-winded       >>>> diatribes.       >>>>       >>>> “We were arguing a lot… we would be up all night, and I would assert       >>>> a boundary, or say, like, ‘I don’t want to have this discussion in       >>>> front of the kids,’ or ‘I need to go to bed,’” she recounted,       “and       >>>> [my ex]       >>>> would immediately turn on ChatGPT and start talking to it, and be       >>>> like, ‘can you believe what she’s doing?'”       >>>>       >>>> Her ex would carry out these conversations with ChatGPT on speaker       >>>> phone, she added — within earshot, pointedly, so she could hear       >>>> everything.       >>>>       >>>> “[My ex] would have it on speaker phone, and then have it speak not       >>>> to me, but it would be in the same room,” she recalled. “And of       >>>> course, ChatGPT was this confirmative voice, being like, ‘you’re so       >>>> right.'”       >>>>       >>>> Today, the former couple, together nearly 15 years, is in the midst       >>>> of a contentious divorce and custody battle.       >>>       >>>> Even Geoffrey Hinton, a Nobel Prize-winning computer scientist known       >>>> as a “Godfather of AI” — a technology that likely wouldn’t exist       in       >>>> its current form without his contributions — recently conceded that       >>>> his girlfriend had broken up with him using ChatGPT.       >>>>       >>>> “She got ChatGPT to tell me what a rat I was… she got the chatbot to       >>>> explain how awful my behavior was and gave it to me,” Hinton told The       >>>> Financial Times. “I didn’t think I had been a rat, so it didn’t make       >>>> me feel too bad.”       >>>       >>> and many more examples of LLM induced retardation @       >>> https://futurism.com/chatgpt-marriages-divorces       >>>       >>> We are currently careening towards a future where people will become       >>> increasingly solitary because some autocomplete trained on billions of       >>> reddit comments is telling them to terminate their relationships.       >>> This is clearly happening on accident, I don't think Sam Altman       >>> personally wants people breaking up because of his chatbot.       >>> However, now that Pandora's box is opened, do you think that       >>> Generative AI companies will begin manipulating people towards       >>> atomization?       >>> Love and the fulfillment derived from it is one of the few things that       >>> cannot be bought and sold in this increasingly financialized present.       >>> There are powerful interests against people deriving enjoyment out of       >>> something that isn't a product.       >>       >>       >> It's the people not the LLM... Why one should ask certain stuff to AI       >> is beyond me       >>       >>       >>       > The people wouldn't act like this if it weren't for the existence of       > LLMs, not at this scale. Many people now admit that they use ChatGPT as       > a therapist or interlocutor. What happens when       > OpenAI/Google/Meta/Anthropic figure out that they can manipulate people       > by modifying the Language models? How are they going to choose to       > manipulate them? They figured out mass manipulation through social media       > quickly enough.              They are comfortably selling themselves. You guys in the states really gave       yourself away.                            --       ---/g/ on Usenet? Madness!---              --- SoupGate-Win32 v1.05        * Origin: you cannot sedate... all the things you hate (1:229/2)    |
[   << oldest   |   < older   |   list   |   newer >   |   newest >>   ]
(c) 1994, bbs@darkrealms.ca