Jump to content

'Catastrophic overtraining' could harm large language AI models that are trained on more data for the sake of training

Featured Replies

Posted

cuJ2nHdA2cLngX4bhsHsye.jpg


  • Researchers from top US universities warn extending pre-training can be detrimental to performance

  • Too much pre-training can deliver worse performance due to something akin to the butterfly effect

  • The more they are pre-trained, the more they become sensitive to small changes that could disrupt the end result


Researchers from Carnegie Mellon, Stanford, Harvard, and Princeton are challenging one of AI development’s accepted core beliefs - that the more pre-training data the better the performance…

View the full article

  • James changed the title to 'Catastrophic overtraining' could harm large language AI models that are trained on more data for the sake of training

Join the conversation

You can post now and register later. If you have an account, sign in now to post with your account.

Guest
Reply to this topic...