Averting model collapse, where artificial intelligence models train on synthetic data and eventually output hallucinations

https://www.eurekalert.org/news-releases/1127473

"synthetic, machine-generated datasets creating closed loop, overcome mapping closed-loop learning behaviors within statistical exponential families, isolating mathematical mechanics causing this... introducing as little as single data point from outside world into closed loop completely prevents AI from hallucinating... anchoring training with established prior knowledge safeguards model stability even when the volume of synthetic data is infinitely larger"

Comments