Forums before death by AOL, social media and spammers... "We can't have nice things"
|    alt.buddha.short.fat.guy    |    Uhhh not sure, something about Buddhism    |    155,846 messages    |
[   << oldest   |   < older   |   list   |   newer >   |   newest >>   ]
|    Message 155,446 of 155,846    |
|    dart200 to Wilson    |
|    Re: Something Big Is Happening (1/3)    |
|    18 Feb 26 16:53:30    |
      From: user7160@newsgrouper.org.invalid              lol yes and i still need to talk to a human for even basic customer       service resolutions 🤷              anyways, dude who bolted a front end onto gpt is writing a hype article              not really worthy news for this sub              On 2/18/26 10:52 AM, Wilson wrote:       > By Matt Shumer • Feb 9, 2026       >       > Think back to February 2020. If you were paying close attention, you       > might have noticed a few people talking about a virus spreading       > overseas. But most of us weren't paying close attention. The stock       > market was doing great, your kids were in school, you were going to       > restaurants and shaking hands and planning trips. If someone told you       > they were stockpiling toilet paper you would have thought they'd been       > spending too much time on a weird corner of the internet. Then, over the       > course of about three weeks, the entire world changed. Your office       > closed, your kids came home, and life rearranged itself into something       > you wouldn't have believed if you'd described it to yourself a month       > earlier.       >       > I think we're in the "this seems overblown" phase of something much,       > much bigger than Covid.       >       > I've spent six years building an AI startup and investing in the space.       > I live in this world. And I'm writing this for the people in my life who       > don't... my family, my friends, the people I care about who keep asking       > me "so what's the deal with AI?" and getting an answer that doesn't do       > justice to what's actually happening. I keep giving them the polite       > version. The cocktail-party version. Because the honest version sounds       > like I've lost my mind. And for a while, I told myself that was a good       > enough reason to keep what's truly happening to myself. But the gap       > between what I've been saying and what is actually happening has gotten       > far too big. The people I care about deserve to hear what is coming,       > even if it sounds crazy.       >       > I should be clear about something up front: even though I work in AI, I       > have almost no influence over what's about to happen, and neither does       > the vast majority of the industry. The future is being shaped by a       > remarkably small number of people: a few hundred researchers at a       > handful of companies... OpenAI, Anthropic, Google DeepMind, and a few       > others. A single training run, managed by a small team over a few       > months, can produce an AI system that shifts the entire trajectory of       > the technology. Most of us who work in AI are building on top of       > foundations we didn't lay. We're watching this unfold the same as you...       > we just happen to be close enough to feel the ground shake first.       >       >       > For years, AI had been improving steadily. Big jumps here and there, but       > each big jump was spaced out enough that you could absorb them as they       > came. Then in 2025, new techniques for building these models unlocked a       > much faster pace of progress. And then it got even faster. And then       > faster again. Each new model wasn't just better than the last... it was       > better by a wider margin, and the time between new model releases was       > shorter. I was using AI more and more, going back and forth with it less       > and less, watching it handle things I used to think required my expertise.       >       > Then, on February 5th, two major AI labs released new models on the same       > day: GPT-5.3 Codex from OpenAI, and Opus 4.6 from Anthropic (the makers       > of Claude, one of the main competitors to ChatGPT). And something       > clicked. Not like a light switch... more like the moment you realize the       > water has been rising around you and is now at your chest.       >       > I am no longer needed for the actual technical work of my job. I       > describe what I want built, in plain English, and it just... appears.       > Not a rough draft I need to fix. The finished thing. I tell the AI what       > I want, walk away from my computer for four hours, and come back to find       > the work done. Done well, done better than I would have done it myself,       > with no corrections needed. A couple of months ago, I was going back and       > forth with the AI, guiding it, making edits. Now I just describe the       > outcome and leave.       >       > Let me give you an example so you can understand what this actually       > looks like in practice. I'll tell the AI: "I want to build this app.       > Here's what it should do, here's roughly what it should look like.       > Figure out the user flow, the design, all of it." And it does. It writes       > tens of thousands of lines of code. Then, and this is the part that       > would have been unthinkable a year ago, it opens the app itself. It       > clicks through the buttons. It tests the features. It uses the app the       > way a person would. If it doesn't like how something looks or feels, it       > goes back and changes it, on its own. It iterates, like a developer       > would, fixing and refining until it's satisfied. Only once it has       > decided the app meets its own standards does it come back to me and say:       > "It's ready for you to test." And when I test it, it's usually perfect.       >       > I'm not exaggerating. That is what my Monday looked like this week.       >       > But it was the model that was released last week (GPT-5.3 Codex) that       > shook me the most. It wasn't just executing my instructions. It was       > making intelligent decisions. It had something that felt, for the first       > time, like judgment. Like taste. The inexplicable sense of knowing what       > the right call is that people always said AI would never have. This       > model has it, or something close enough that the distinction is starting       > not to matter.       >       > I've always been early to adopt AI tools. But the last few months have       > shocked me. These new AI models aren't incremental improvements. This is       > a different thing entirely.       >       > And here's why this matters to you, even if you don't work in tech.       >       > The AI labs made a deliberate choice. They focused on making AI great at       > writing code first... because building AI requires a lot of code. If AI       > can write that code, it can help build the next version of itself. A       > smarter version, which writes better code, which builds an even smarter       > version. Making AI great at coding was the strategy that unlocks       > everything else. That's why they did it first. My job started changing       > before yours not because they were targeting software engineers... it       > was just a side effect of where they chose to aim first.       >       > They've now done it. And they're moving on to everything else.       >       > The experience that tech workers have had over the past year, of       > watching AI go from "helpful tool" to "does my job better than I do", is       > the experience everyone else is about to have. Law, finance, medicine,       > accounting, consulting, writing, design, analysis, customer service. Not              [continued in next message]              --- SoupGate-Win32 v1.05        * Origin: you cannot sedate... all the things you hate (1:229/2)    |
[   << oldest   |   < older   |   list   |   newer >   |   newest >>   ]
(c) 1994, bbs@darkrealms.ca