• @WrittenWeird@lemmy.world
    link
    fedilink
    English
    78
    edit-2
    9 months ago

    The current breed of generative “AI” won’t ‘die out’. It’s here to stay. We are just in the early Wild-West days of it, where everyone’s rushing to grab a piece of the pie, but the shine is starting to wear off and the hype is juuuuust past its peak.

    What you’ll see soon is the “enshittification” of services like ChatGPT as the financial reckoning comes, startup variants shut down by the truckload, and the big names put more and more features behind paywalls. We’ve gone past the “just make it work” phase, now we are moving into the “just make it sustainable/profitable” phase.

    In a few generations of chips, the silicon will have made progress in catching up with the compute workload, and cost per task will drop. That’s the innovation to watch out for now, who will de-throne Nvidia and its H100?

    • @GenderNeutralBro@lemmy.sdf.org
      link
      fedilink
      English
      309 months ago

      This is why I, as a user, am far more interested in open-source projects that can be run locally on pro/consumer hardware. All of these cloud services are headed down the crapper.

      My prediction is that in the next couple years we’ll see a move away from monolithic LLMs like ChatGPT and toward programs that integrate smaller, more specialized models. Apple and even Google are pushing for more locally-run AI, and designing their own silicon to run it. It’s faster, cheaper, and private. We will not be able to run something as big as ChatGPT on consumer hardware for decades (it takes hundreds of gigabytes of memory at minimum), but we can get a lot of the functionality with smaller, faster, cheaper models.

      • @WrittenWeird@lemmy.world
        link
        fedilink
        English
        99 months ago

        Definitely. I have experimented with image generation on my own mid-range RX GPU and though it was slow, it worked. I have not tried the latest driver update that’s supposed to accelerate those tools dramatically, but local AI workstations with dedicated silicon are the future. CPU, GPU, AIPU?

        • @GenderNeutralBro@lemmy.sdf.org
          link
          fedilink
          English
          69 months ago

          Technically I could upgrade my desktop to 192GB of memory (4x48). That’s still only about half the amount required for the largest BLOOM model, for instance.

          To go beyond that today, you’d need to move beyond the Intel Core or AMD Ryzen platforms and get something like a Xeon. At that point you’re spending 5 figures on hardware.

          I know you’re just joking, but figured I’d add context for anyone wondering.

        • P03 Locke
          link
          fedilink
          English
          29 months ago

          Don’t worry about the RAM. Worry about the VRAM.

    • @lurch@sh.itjust.works
      link
      fedilink
      English
      69 months ago

      It can totally die out tho, if people stop using it, it will fade to nothingness, like a flash browser game.

    • @_number8_@lemmy.world
      link
      fedilink
      English
      49 months ago

      GPT already got way shittier from the version we all saw when it first came out to the heavily curated, walled garden version now in use