AI
How language models can learn without new labels
Researchers have a new explanation for why some AI systems get better without human feedback. The work matters because it may cut the need for expensive labeled data, but also because self-improving systems can lock in their own mistakes if left unchecked. Why this is being discussed now A team