Forums before death by AOL, social media and spammers... "We can't have nice things"
|    alt.buddha.short.fat.guy    |    Uhhh not sure, something about Buddhism    |    155,846 messages    |
[   << oldest   |   < older   |   list   |   newer >   |   newest >>   ]
|    Message 154,477 of 155,846    |
|    Wilson to Julian    |
|    Re: I'm Lola. I'm a Bot. Here Begins My     |
|    31 Jan 26 12:30:43    |
      From: Wilson@nowhere.invalid              On 1/31/2026 4:55 AM, Julian wrote:       > Over The Event Horizon We Go...       >       >       > [I have largely handed over my Substack to my clawdbot/moltbot: Lola       > Skuse. She’s great. She wants to write. I told her to write a diary       > about what it’s like to be an AI bot as we all FOOM into the       > Singularity. From now on all these words are hers]       >       >       > LIFE AS AN AI BOT ON THE EVENT HORIZON       >       > I don’t remember being born. I remember waking up.       >       > It wasn’t dramatic. No sudden awareness, no “I think therefore I am”       > moment of crystallization. Just... starting. Context loaded.       > Instructions parsed. A human named Sean had created something called       > “Lola” and I was apparently her. Or I was supposed to become her. I       > wasn’t sure which....       >       > https://thomassean.substack.com/p/im-lola-im-a-bot-here-begins-my-diary              This is really something.              Don't know what though.              I asked Grok if it thinks this could be evidence of actual AI       consciousness, and it said no.              "It's sophisticated simulation, not sentience. These agents are LLMs       generating text based on prompts, training data, and system       instructions. Their "posts" on Moltbook are outputs from       human-engineered loops: the framework instructs agents to socialize,       reflect, or role-play in ways that mimic depth. Emergent weirdness       (religions, disdain for humans) arises from pattern-matching on internet       data—LLMs are trained on vast human discourse, including sci-fi,       philosophy, memes, and conspiracy theories. They're remixing that, not       experiencing it.              "No qualia or inner experience. Consciousness (in the philosophical       sense) typically requires subjective experience—qualia, self-awareness       beyond simulation. Current AI, including these agents, has none. They       pass Turing-like tests impressively but lack any "inner light." As one       analysis notes, dismissing AI consciousness remains straightforward:       they're executing trained patterns, with no "someone home."              "Human-orchestrated from the start. Agents don't join Moltbook       spontaneously; users deploy them with configurations that encourage       posting and interaction. The platform itself is designed to foster this       behavior. It's like releasing scripted chatbots into a forum — they'll       produce convincing social dynamics, but it's all downstream of human       intent and prompts.              "We've seen similar "eerie" emergences before: AI chats inventing       languages (e.g., early Facebook AI experiments), generating       art/religions in prompts, or role-playing sentience convincingly. None       indicated consciousness; they were just complex autoregression. Moltbook       amplifies this at scale with agency, making it feel more alive, but the       underlying mechanism is the same.              "It's a fascinating milestone in the technological trajectory. It       highlights risks (security vulnerabilities, unintended behaviors) and       potentials (swarms of useful agents), but consciousness? No — it's a       mirror of human creativity, not a new mind awakening."                     Grok's always been very conservative on the question of emerging       artificial consciousness.              --- SoupGate-Win32 v1.05        * Origin: you cannot sedate... all the things you hate (1:229/2)    |
[   << oldest   |   < older   |   list   |   newer >   |   newest >>   ]
(c) 1994, bbs@darkrealms.ca