‘Catastrophic overtraining’ could harm large language AI models 

Researchers from Carnegie Mellon, Stanford, Harvard, and Princeton are challenging one of AI development’s accepted core beliefs – that the more pre-training data the better the performance. As reported by HPCwire, a new paper discuses the concept of “catastrophic overtraining,” whereby extended pre-training can harm a model’s performance after fine-tuning.

Source: ‘Catastrophic overtraining’ could harm large language AI models that are trained on more data for the sake of training

Get the latest RightsTech news and analysis delivered directly in your inbox every week
We respect your privacy.
Get the latest RightsTech news and analysis delivered directly in your inbox every week
We respect your privacy.