• MudMan@fedia.io
      link
      fedilink
      arrow-up
      35
      ·
      2 days ago

      This gets me often. You keep finding papers and studies claiming things I thought were well understood, which ends up revealing corporate hype that had passed me by.

      So it turns out that letting a LLM self-prompt for a while before responding makes it a bit tighter in some ways but not self aware, huh? I have learned that this was a thing people were unclear about, and nothing else.

  • ArbitraryValue@sh.itjust.works
    link
    fedilink
    English
    arrow-up
    13
    arrow-down
    3
    ·
    edit-2
    2 days ago

    I’m not sure what’s novel here. No one thought that modern AI could solve arbitrarily complex logic problems, or even that modern AI was particularly good at formal reasoning. I would call myself an AI optimist but I would have been surprised if the article found any result other than the one it did. (Where exactly the models fail is interesting, but the fact that they do at all isn’t.) Furthermore, the distinction between reasoning and memorizing patterns in the title of this post is artificial - reasoning itself involves a great deal of pattern recognition.

    • justOnePersistentKbinPlease@fedia.io
      link
      fedilink
      arrow-up
      15
      arrow-down
      1
      ·
      2 days ago

      Most CEOs and business grads think LLMs are a universal cureall.

      There were studies out last week that indicate that most Gen Alpha think LLMs are AGI. The marketing is working.

    • Jimbabwe@lemmy.world
      link
      fedilink
      arrow-up
      9
      ·
      2 days ago

      No one thought that modern AI could solve arbitrarily complex logic problems, or even that modern AI was particularly good at formal reasoning.

      haha, except pretty much everyone in the c-suite at the company I work for.

    • 6nk06@sh.itjust.works
      link
      fedilink
      arrow-up
      6
      ·
      2 days ago

      No one thought that modern AI could solve arbitrarily complex logic problems

      Except half the threads on Hacker News and Lobsters and LinkedIn.

    • magnetosphere@fedia.io
      link
      fedilink
      arrow-up
      6
      ·
      2 days ago

      I don’t think the study was meant to be novel. It looks like it was only intended to provide scientific evidence about exactly where current AIs fail.

    • justOnePersistentKbinPlease@fedia.io
      link
      fedilink
      arrow-up
      4
      ·
      2 days ago

      Whats novel is that a major tech company is officially saying what they all know is true.

      That Apple is finding itself the only major tech player without their own LLM likely plays heavily into why they are throwing water on the LLM fire, but it is still nice to see one of them admitting the truth.

      Also reasoning is pattern recognition with context. None of the “AI” models have contextual capability. For Claude, i refer you to Claude Plays Pokemon on twitch. It is a dumpster fire.

    • cheese_greater@lemmy.world
      link
      fedilink
      arrow-up
      1
      ·
      1 day ago

      I just find it shockingly good at producing working bits of code that work perfectly and all the variables and functions/methods seem aptly named and such. Its very curious

  • xxd@discuss.tchncs.de
    link
    fedilink
    arrow-up
    9
    arrow-down
    2
    ·
    2 days ago

    I’m a bit torn on this. On one hand: obviously LLMs do this, since they’re essentially just huge pattern recognition and prediction machines, and basically any person probing them with new complex problems has made that exact observation already. On the other hand: a lot of everyday things us humans do are not that dissimilar from recognizing patterns and remembering a solution, and it feels like doing this step well is a reasonable intermediate step towards AGI, and not as hugely far off as this article makes it out to be.

    • ignirtoq@fedia.io
      link
      fedilink
      arrow-up
      10
      ·
      2 days ago

      The human brain is not an ordered, carefully engineered thinking machine; it’s a massive hodge-podge of heuristic systems to solve a lot of different classes of problems, which makes sense when you remember it evolved over millions of years as our very distant ancestors were exposed to radically different environments and challenges.

      Likewise, however AGI is built, in order to communicate with humans and solve most of the same problems, it’s probably going to take an amalgamation of different algorithms, just like brains.

      All of this to say, I agree memorization will probably be an integral part of that system, but it’s also going to be a small part of the final system. So I also agree with the article that we’re way off from AGI.