Strong hallucinations from negation and how to fix them
Published in Findings of ACL 2024, 2024
Despite great performance on many tasks, language models (LMs) still struggle with reasoning, sometimes providing responses that cannot possibly be true because they stem from logical incoherence. We call such responses strong hallucinations and prove that they follow from its computation of its internal representations for logical operators and outputs from those representations Read more
Recommended citation: Asher, Nicholas, and Swarnadeep Bhar. "Strong hallucinations from negation and how to fix them." arXiv preprint arXiv:2402.10543 (2024).