ChatGPT has meltdown and starts sending alarming messages to users::AI system has started speaking nonsense, talking Spanglish without prompting, and worrying users by suggesting it is in the room with them

  • @thehatfox@lemmy.world
    link
    fedilink
    English
    4910 months ago

    The development of LLMs is possibly becoming self defeating, because the training data is being filled not just with human garbage, but also AI garbage from previous, cruder LLMs.

    We may well end up with a machine learning equivalent of Kessler syndrome, with our pool of available knowledge eventually becoming too full of junk to progress.

    • @CarbonIceDragon@pawb.social
      link
      fedilink
      English
      1910 months ago

      I mean, surely the solution to that would be to use curated/vetted training data? Or at the very least, data from before LLMs became commonplace?

      • @KevonLooney@lemm.ee
        link
        fedilink
        English
        1910 months ago

        The funny thing is, children are similar. They just learn whatever you put in front of them. We have whole systems for educating children for decades of their lives.

        With AI we literally just plopped them in front of the Internet, with no guidelines on what to learn. AI researchers say “it’s a black box! We don’t know why it’s doing this!” You fed it everything you could and gave it few rules on what to do. You are the reason why it’s nuts.

        Humans come hardwired to be a certain way, do certain things. Maybe they need to start AI off like that, some basic programs that guide learning. “Learn everything” isn’t working.

        • @thehatfox@lemmy.world
          link
          fedilink
          English
          810 months ago

          Humans come hardwired to be a certain way, do certain things. Maybe they need to start AI off like that, some basic programs that guide learning. “Learn everything” isn’t working.

          That’s a good point. For real brains, size and intelligence are not linked. An elephant brain has 3 times the amount of neurons as a human brain, but a human brain is more intelligent. There is more to intelligence than just the amount of neutrons, real or virtual, so making larger and larger AI models may not be the right direction.

          • @KevonLooney@lemm.ee
            link
            fedilink
            English
            510 months ago

            True. Maybe they just need more error correction. Like spend more energy questioning whether what you say is true. Right now LLMs seems to just vomit out whatever they thought up, with no consideration of whether it makes sense.

            They’re like an annoying friend who just can’t shut up.

            • @nilloc@discuss.tchncs.de
              link
              fedilink
              English
              210 months ago

              They aren’t thinking though. They’re making connection with the trained data that they’ve processed.

              This is really clear when they are asked to write code worth to vague a prompt.

              Maybe feeding them through primary school curriculum (including essays and tests) would be helpful, but I don’t think the language models really sort knowledge yet.

      • Ms. ArmoredThirteen
        link
        fedilink
        English
        910 months ago

        Yes but that only works if we can differentiate that data on a pretty big scale. The only way I can see it working at scale is by having meta data to declare if something is AI generated or not. But then we’re relying on self reporting so a lot of people have to get on board with it and bad actors can poison the data anyway. Another way could be to hire humans to chatter about specific things you want to train it on which could guarantee better data but be quite expensive. Only training on data from before LLMs will turn it into an old people pretty quickly and it will be noticable when it doesn’t know pop culture or modern slang.

        • @5too@lemmy.world
          link
          fedilink
          English
          510 months ago

          Pretty sure this is why they keep training it on books, movies, etc. - it’s already intended to make sense, so it doesn’t need curated.

    • @Asafum@feddit.nl
      link
      fedilink
      English
      1310 months ago

      God I hope all those CEOs and greedy fuckheads that fired hundreds of thousands of people wayyyyy too soon to replace them with this get their pants shredded by the fallout.

      Naturally they’ll get their golden parachutes and land on their feet even richer than before, but it’s nice to dream lol

    • Ms. ArmoredThirteen
      link
      fedilink
      English
      810 months ago

      This is called model collapse and imo has to be solved if LLMs are to be a long term thing. I could see it wrecking this current AI push until people step back and reevaluate how data gets sucked up

    • @nexusband@lemmy.world
      link
      fedilink
      English
      710 months ago

      I really hope so. I still have to see a meaningful use case for these kind of LLMs that just get fed with all kinds of data. LLMs “on premise” that are used for specific jobs are fine, but this…I really hope a Kessler-Like syndrome blows it out the water, for countless reasons…

    • @kent_eh@lemmy.ca
      link
      fedilink
      English
      210 months ago

      but also AI garbage from previous, cruder LLMs

      And now I’m picturing it training on a bunch of chats with Eliza…

    • @Paragone@lemmy.world
      link
      fedilink
      English
      -110 months ago

      Damn.

      Thank you VERY much for that insight: AI’s version of Kessler-syndrome.

      EXACTLY.

      Damn, damn, damn, that gets the truth right in its marrow.

      _ /\ _