International Journal of Innovative Research in Computer and Communication Engineering

ISSN Approved Journal | Impact factor: 8.771 | ESTD: 2013 | Follows UGC CARE Journal Norms and Guidelines

| Monthly, Peer-Reviewed, Refereed, Scholarly, Multidisciplinary and Open Access Journal | High Impact Factor 8.771 (Calculated by Google Scholar and Semantic Scholar | AI-Powered Research Tool | Indexing in all Major Database & Metadata, Citation Generator | Digital Object Identifier (DOI) |


TITLE Addressing Hallucination in Large Language Models Through a Hybrid Detection-Verification Retrieval-Augmented Generation Framework
ABSTRACT Advances in transformer-based language modeling have enabled machines to produce remarkably coherent and contextually relevant text across a wide range of tasks. Yet, these powerful systems carry a persistent flaw: they sometimes generate content that is plausible in tone but factually groundless — a phenomenon commonly referred to as hallucination. This shortcoming becomes particularly consequential in high-stakes domains such as clinical decision support, regulatory compliance, and legal advisory systems, where even minor inaccuracies can have serious downstream effects. Retrieval-Augmented Generation (RAG) has garnered considerable attention as a strategy to anchor language model outputs in externally verified knowledge, thereby mitigating the tendency toward fabrication. However, practical deployments of RAG expose a set of underexplored vulnerabilities: retrieved documents may introduce noise, generated content may conflict with retrieved evidence, and the pipeline often lacks a formal mechanism to detect or correct factual inconsistencies post-generation. This paper systematically examines five prominent research contributions addressing hallucination within the RAG paradigm and identifies recurring structural limitations across them. Building on this analysis, we introduce the Hybrid Detection-Verification RAG (HDV-RAG) framework — a multi-stage architecture that brings together semantic retrieval, cross-encoder re-ranking, hallucination classification with explanatory outputs, knowledge cross-verification, and Direct Preference Optimization (DPO)-driven fine-tuning into a single coherent pipeline. Empirical evaluation across multiple benchmarks demonstrates that HDV-RAG achieves measurable gains in factual consistency, response reliability, and BERTScore compared to standard RAG configurations.
AUTHOR ANUSHA K, BHARGAVI NAIK, BHAVANA MOHANDAS SANGALAD, PROF. USHA K Students, Department of Computer Science and Engineering, Jain Institute of Technology, Davangere, Karnataka, India Professor, Department of Computer Science and Engineering, Jain Institute of Technology, Davangere, Karnataka, India
VOLUME 184
DOI DOI: 10.15680/IJIRCCE.2026.1405015
PDF pdf/15_Addressing Hallucination in Large Language Models Through a Hybrid Detection-Verification Retrieval-Augmented Generation Framework.pdf
KEYWORDS
References [1] J. Song, X. Wang, J. Zhu, Y. Wu, X. Cheng, R. Zhong, and C. Niu, "RAG-HAT: A Hallucination-Aware Tuning Pipeline for LLM in Retrieval-Augmented Generation," in Proc. 2024 Conference on Empirical Methods in Natural Language Processing (EMNLP Industry Track), Miami, USA, Nov. 2024, pp. 1548–1558.
[2] P. Béchard and O. Marquez Ayala, "Reducing Hallucination in Structured Outputs via Retrieval-Augmented Generation," in Proc. NAACL-HLT 2024 (Industry Track), 2024, pp. 228–238.
[3] M. Y. Mohammed, S. A. Ali, S. K. Ali, A. A. Majeed, and E. H. Mohamed, "Aftina: Enhancing Stability and Preventing Hallucination in AI-Based Islamic Fatwa Generation Using LLMs and RAG," Neural Computing and Applications, vol. 37, pp. 20957–20982, 2025.
[4] J. Kirchenbauer and C. Barns, "Hallucination Reduction in Large Language Models with Retrieval-Augmented Generation Using Wikipedia Knowledge," 2024.
[5] P. Béchard and O. Marquez Ayala, "Reducing Hallucination in Structured Outputs via Retrieval-Augmented Generation," ServiceNow Research, 2024.
image
Copyright © IJIRCCE 2020.All right reserved