A new paper suggests diminishing returns from larger and larger generative AI models. Dr Mike Pound discusses.

The Paper (No “Zero-Shot” Without Exponential Data): https://arxiv.org/abs/2404.04125

  • @[email protected]
    link
    fedilink
    English
    2
    edit-2
    7 months ago

    no matter our computational techniques, the diminishing returns in predictive precision is reached far sooner than we achieve general intelligence

    That’s very bold presumption. How can they be so sure of this, that any future models can’t tackle the issue? have they got proof or something.

    • @[email protected]
      link
      fedilink
      English
      27 months ago

      No, they just calculate with increased size of the training roster… it’s not that complicated. Which is a fair presumption as that is how we’ve increased the predictive precision so far.

    • @[email protected]
      link
      fedilink
      English
      1
      edit-2
      7 months ago

      It seems far more bold to presume that general intelligence will be created any time soon when current machine learning is nowhere close.