The argument for current LLM AIs leading to AGI has always been that they would spontaneously develop independent reasoning, through an unknown emergent property that would appear as they scale. It hasn’t happened, and there’s no sign that it will.

That’s a dilemma for the big AI companies. They are burning through billions of dollars every month, and will need further hundreds of billions to scale further - but for what in return?

Current LLMs can still do a lot. They’ve provided Level 4 self-driving, and seem to be leading to general-purpose robots capable of much useful work. But the headwinds look ominous for the global economy, - tit-for-tat protectionist trade wars, inflation, and a global oil shock due to war with Iran all loom on the horizon for 2025.

If current AI players are about to get wrecked, I doubt it’s the end for AI development. Perhaps it will switch to the areas that can actually make money - like Level 4 vehicles and robotics.

  • drspod
    link
    fedilink
    English
    arrow-up
    14
    arrow-down
    1
    ·
    1 month ago

    Current LLMs can still do a lot. They’ve provided Level 4 self-driving, and seem to be leading to general-purpose robots capable of much useful work.

    Really? I don’t think this has anything to do with LLMs. They are likely using reinforcement learning combined with traditional AI techniques, an approach which has been the foundation of these kinds of robotics and automation for decades at this point.

    If other areas of AI and automation have seen a boost at the same time as LLMs came on the scene, it’s because the underlying hardware has become so much faster, cheaper and easily available, along with the massively increased interest in and funding for these types of research, and computer scientists re-skilling into a discipline that’s in the midst of a bubble.

    • TechLich@lemmy.world
      link
      fedilink
      English
      arrow-up
      3
      ·
      1 month ago

      Not entirely true, the big change was multi-headed attention and the transformer model.

      It’s not just being used for language but anything where sequence and context patterns are really important. Some stuff is still using convolutional networks and RNNs etc. but transformers aren’t just for LLMs. There’s definitely a lot of algorithmic advances driving the wave of new ai implementations, not just hardware improvements.

      • drspod
        link
        fedilink
        English
        arrow-up
        3
        ·
        1 month ago

        Thanks for the clarification. The point remains that it’s not true to say that LLMs have “provided Level 4 self-driving and … general-purpose robots.”

        • TechLich@lemmy.world
          link
          fedilink
          English
          arrow-up
          2
          ·
          1 month ago

          Agreed. It’s a lot of the same tech that powers both, but it’s not like a self driving car contains a language model that’s fine tuned on the adventures of Steve MacQueen or something.