PHPFND: Detecting Fake News via Post-Hoc Processing of LLMs Hallucination
DOI:
https://doi.org/10.1609/aaai.v40i1.37050Abstract
Large Language Models (LLMs) perform excellently in fake news detection tasks, but their outputs are often accompanied by hallucinations, i.e., generated content that is contradictory to facts. Previous studies have mostly mitigated hallucinations through prompt design. However, this paper reveals that regions in news articles which easily induce hallucinations in LLMs correspond closely to the most challenging regions for fake news detectors. In this paper, we propose a fake news detection framework (PHPFND) based on post-hoc processing of LLMs hallucination. Specifically, our framework includes a hallucination detection module (ISHD) based on information structuring that detects three types of hallucinations in LLMs in a targeted manner, and a hallucination-driven feature enhancement mechanism (HDFE) that incorporates hallucination signals as explicit features into sentence-level encoding and feature fusion to guide the model’s attention toward high-risk regions. Experimental results on two mainstream fake news datasets show that our proposed method significantly outperforms LLM-based baselines.Downloads
Published
2026-03-14
How to Cite
Ma, J., Ma, J., Zhang, W., & Liu, Y. (2026). PHPFND: Detecting Fake News via Post-Hoc Processing of LLMs Hallucination. Proceedings of the AAAI Conference on Artificial Intelligence, 40(1), 827-835. https://doi.org/10.1609/aaai.v40i1.37050
Issue
Section
AAAI Technical Track on Application Domains I