Research on Hallucination Mitigation in Large Language Models
Main Article Content
Keywords
large language models, hallucination mitigation, multimodality
Abstract
The proliferation of large language models has brought the issue of hallucinations, which are outputs that deviate from inputs or fabricate information, to the forefront of concerns regarding reliability and deployment safety. This paper provides a systematic review of existing mitigation strategies and evaluation frameworks for hallucinations. Among mitigation approaches, supervised fine-tuning and reinforcement learning from human feedback have demonstrated moderate success; however, their heavy reliance on extensive, high-quality human annotations renders them costly and difficult to scale. On the evaluation side, this study examines several multimodal benchmarks, including POPE, MMHal-Bench, MM-Vet, and MMBench, which enable comprehensive assessment of model outputs through a blend of automated metrics and human judgment across multiple dimensions. Overall, substantial progress has been achieved in hallucination evaluation frameworks, yet scalable and cost-effective mitigation techniques remain elusive, particularly for hallucinations in complex multimodal settings, where further breakthroughs are urgently needed.
References
- Li, Y., Du, Y., Zhou, K., Wang, J., Zhao, X. and Wen, J.-R., Published. Evaluating Object Hallucination in Large Vision-Language Models. The 2023 Conference on Empirical Methods in Natural Language Processing, Singapore. Optica Publishing Group, pp. 292-305.
- Li, Y. Q., Wang, X. W. and Wang, Q. Y., (2025). Research on multimodal sentiment analysis models for social media based on multimodal large language models. Information Studies:Theory & Application, vol. 48, no. 11, pp. 188-197.
- Lin, Y. W., (2025). Research on large model hal lucination mitigation methods based on contrastive decoding. Master's Theses, Guangzhou University.
- Sun, Z., Shen, S., Cao, S., Liu, H., Li, C., Shen, Y., Gan, C., Gui, L., Wang, Y.-X. and Yang, Y., (2024). Published. Aligning large multimodal models with factually augmented rlhf. Findings of the Association for Computational Linguistics: ACL 2024, Bangkok, Thailand. Association for Computational Linguistics (ACL), pp. 13088-13110.
- Zhang, J. Q., (2025). Research on hallucination mitigation in multi-modal large language models. Master's Theses, University of Electronic Science and Technology of China.
- Zheng, H. J., (2025). From ChatGPT to Deep Seek: A regulatory approach to the risk of illusion in large models [Online]. Journal of Intelligence. Available: http://kns.cnki.net/kcms/detail/61.1167.g3.20250909.1241.002.html.
