Abstract
Ambiguity in Large Language Models (LLMs) serves as a primary pathway to hallucination generation, undermining trust in AI systems across critical applications. This survey provides the first comprehensive analysis linking ambiguity processing to hallucination mechanisms, establishing how unresolved linguistic, knowledge-based, pragmatic, and cross-modal ambiguities create pathways to fabricated information. Through systematic examination of theoretical foundations, detection methodologies, and resolution approaches, we reveal that current LLMs process ambiguity through constraint satisfaction mechanisms, yet exhibit poor correlation between internal representations and human disambiguation strategies. Analysis of thirty specialized benchmarks demonstrates that while sophisticated resolution methods show promise, from strategic prompting to multi-agent collaboration, fundamental limitations persist in computational scalability, cultural adaptation, and real-world deployment. The findings identify four critical research frontiers: developing architectures with native uncertainty representation, bridging scalability gaps between laboratory success and deployment constraints, creating culturally-aware disambiguation strategies, and establishing verification frameworks for genuine understanding. These challenges represent fundamental tests of AI system intelligence that will determine whether future LLMs can navigate linguistic uncertainty effectively or remain constrained by brittle pattern matching.
| Original language | English |
|---|---|
| Journal | Natural Language Processing Journal |
| DOIs | |
| Publication status | E-pub ahead of print - 14 Jul 2025 |
Bibliographical note
The author, Prof XIE Haoran, is an Editor-in-Chief for this journal and was not involved in the editorial review or the decision to publish this article.Funding
This work has been supported by a grant from Hong Kong Metropolitan University (Project Reference No. CP/2022/02) and by the Hong Kong Research Grants Council through the Faculty Development Scheme (Project No. UGC/FDS16/E10/23).
Keywords
- Ambiguity
- LLM
- Hallucination
- Retrieval-augmented generation
- Multi-agent systems