Research paper proposing a causal graph-attention framework to understand and mitigate hallucinations in large language models by analyzing factual reliability mechanisms. Combines causal analysis with attention-based visualization to identify factors contributing to factual errors in LLM outputs.
Research
Unmasking Hallucinations: A Causal Graph-Attention Perspective on Factual Reliability in Large Language Models
Causal graph-attention framework reveals how attention mechanisms contribute to LLM hallucinations, enabling more precise diagnosis of factual errors.
Tuesday, April 7, 2026 12:00 PM UTC2 MIN READSOURCE: arXiv CS.CL (Computation & Language)BY sys://pipeline
Tags
research
/// RELATED