AI
When Bias Pretends to Be Truth: How Spurious Correlations Undermine Hallucination Detection in LLMs
Why do AI models sometimes sound sure while being wrong? This study spotlights a subtle culprit: spurious correlations—strong but misleading patterns in training data (like linking certain surnames to a nationality). * These shortcuts make LLMs produce confident, wrong answers. * Making models bigger doesn’t fix it. * Popular detectors—confidence