I’m usually the one saying “AI is already as good as it’s gonna get, for a long while.”

This article, in contrast, is quotes from folks making the next AI generation - saying the same.

  • Greg Clarke@lemmy.ca
    link
    fedilink
    English
    arrow-up
    5
    arrow-down
    16
    ·
    27 days ago

    Yes of course I’m asserting that. While the performance of LLMs may be plateauing, the cost, context window, and efficiency is still getting much better. When you chat with a modern chat bot it’s not just sending your input to an LLM like the first public version of ChatGPT. Nowadays a single chat bot response may require many LLM requests along with other techniques to mitigate the deficiencies of LLMs. Just ask the free version of ChatGPT a question that requires some calculation and you’ll have a better understanding of what’s going on and the direction of the industry.

    • MajorHavoc@programming.devOP
      link
      fedilink
      English
      arrow-up
      9
      ·
      27 days ago

      I think you’re agreeing, just in a rude and condescending way.

      There’s a lot of ways left to improve, but they’re not as simple as just throwing more data and CPU at the problem, anymore.

      • Greg Clarke@lemmy.ca
        link
        fedilink
        English
        arrow-up
        3
        arrow-down
        3
        ·
        edit-2
        27 days ago

        I’m sorry if I’m coming across as condescending, that’s not my intent. It’s never been “as simple as just throwing more data and CPU at the problem”. There were algorithmic challenges for every LLM evolution. There are still lots of potential improvements using the existing training data. But even if there wasn’t, we’ll still see loads of improvements in chat bots because of other techniques.

        Edit: typo