A new paper suggests diminishing returns from larger and larger generative AI models. Dr Mike Pound discusses.

The Paper (No “Zero-Shot” Without Exponential Data): https://arxiv.org/abs/2404.04125

  • @Murvel
    link
    English
    7
    edit-2
    11 days ago

    What you mentioned is assumed video and paper in question.

    The main argument being that no matter our computational techniques, the diminishing returns in predictive precision is reached far sooner than we achieve general intelligence.

    • @boyi@lemmy.sdf.org
      link
      fedilink
      English
      2
      edit-2
      10 days ago

      no matter our computational techniques, the diminishing returns in predictive precision is reached far sooner than we achieve general intelligence

      That’s very bold presumption. How can they be so sure of this, that any future models can’t tackle the issue? have they got proof or something.

      • @Murvel
        link
        English
        210 days ago

        No, they just calculate with increased size of the training roster… it’s not that complicated. Which is a fair presumption as that is how we’ve increased the predictive precision so far.

      • @technocrit@lemmy.dbzer0.com
        link
        fedilink
        English
        1
        edit-2
        10 days ago

        It seems far more bold to presume that general intelligence will be created any time soon when current machine learning is nowhere close.

    • @Womble@lemmy.world
      link
      fedilink
      English
      210 days ago

      No the argument is current techniques give logarithmic returns in data size, which is bad. But it said nothing about other potential techniques or made any suggestion that this was a general result.

      • @Murvel
        link
        English
        310 days ago

        Well obviously they cannot rule out techniques no one has though of but likewise they obviously accounted for what they deemed to be within the realm of possibility