Skip to content

CiteGuard: Faithful Citation Attribution for LLMs via Retrieval-Augmented Validation

Conference: ACL 2026 arXiv: 2510.17853 Code: https://github.com/KathCYM/CiteGuard Area: AIGC Detection Keywords: citation attribution, retrieval-augmented validation, scientific writing, hallucination mitigation, agent

TL;DR

CiteGuard proposes a retrieval-augmented agent framework with extended retrieval actions (including full-text search and context retrieval) to provide a more faithful basis for scientific citation attribution, achieving 68.1% accuracy on the CiteME benchmark — a 10-point improvement over baselines, approaching human performance (69.2%).

Background & Motivation

Background: LLMs are increasingly used for scientific writing assistance, but citation hallucination is a severe problem (LLMs can generate up to 78-90% fabricated citations). Over 50 citation hallucinations were found among 300 ICLR 2026 submissions.

Limitations of Prior Work: (1) LLM-as-a-Judge has extremely low recall in citation verification (only 16-17%), as LLMs are overly sensitive to minor terminology changes; (2) existing methods like CiteAgent still fall far below human accuracy; (3) existing methods lack the ability to search full-text content of papers.

Key Challenge: Retrieval based solely on titles and abstracts is insufficient to confirm citation relationships; it often requires delving into the full text for cross-validation.

Goal: Design a more faithful and generalizable citation attribution agent.

Key Insight: Extend the set of retrieval actions, particularly by adding full-text search and context retrieval capabilities.

Core Idea: Citation verification requires information beyond the title/abstract level; full-text search and context retrieval provide a stronger evidence base.

Method

Overall Architecture

CiteGuard is an LLM-based agent that extends CiteAgent with three new actions: find_in_text (full-text search within a paper), ask_for_more_context (source paper context retrieval), and search_text_snippet (cross-paper full-text snippet search). It supports iterative retrieval to recommend multiple references.

Key Designs

  1. Extended Retrieval Action Set:

    • Function: Provide deeper evidence than title/abstract
    • Mechanism: Three new actions — find_in_text (search queries within a specific paper's full text), ask_for_more_context (retrieve 3 paragraphs before and after an excerpt's context), and search_text_snippet (cross-database full-text snippet search)
    • Design Motivation: Citation relationships are often hidden in the body of papers; looking only at titles and abstracts may lead to misjudgment
  2. Iterative Multi-Citation Recommendation:

    • Function: Recommend multiple relevant references
    • Mechanism: Each run recommends one reference; subsequent runs exclude already-selected papers to search for new references. Filtering through the exclusion set \(E_k\) ensures no duplicate recommendations
    • Design Motivation: Many academic claims have multiple valid citations; a single reference is insufficient
  3. Cross-Domain Generalization:

    • Function: Evaluate method usability beyond computer science
    • Mechanism: Collect the CiteMulti extended benchmark covering biomedical, physics, and mathematics domains, as well as long-paragraph scenarios
    • Design Motivation: Verify the generality of the method

Loss & Training

No model training involved. The agent uses GPT-4o or DeepSeek-R1 as the base model.

Key Experimental Results

Main Results

CiteME Benchmark Results

Method All-Difficulty Accuracy
CiteAgent + GPT-4o 35.4%
CiteGuard + GPT-4o 45.4% (+10pp)
CiteGuard + DeepSeek-R1 68.1%
Human Performance 69.2%

Ablation Study

  • CiteGuard can identify alternative valid citations not covered by the benchmark
  • The new retrieval actions (especially find_in_text) contribute most to performance improvement
  • Cross-domain experiments show the method has generalization potential

Key Findings

  • Full-text search capability is critical for citation verification
  • The 68.1% accuracy approaching human performance demonstrates the method's effectiveness
  • LLM-as-a-Judge is unreliable for citation verification and requires retrieval augmentation

Highlights & Insights

  • Addresses a real pain point in scientific writing with high practical value
  • Approaching human performance is an important milestone
  • The extended CiteMulti benchmark fills the gap in cross-domain evaluation

Limitations & Future Work

  • Depends on the Semantic Scholar API, which may not cover all domains
  • Full-text search requires paper accessibility; some papers may be unavailable
  • Iterative retrieval increases inference cost
  • Future work could explore integrating the method into academic writing workflows
  • The extension of CiteAgent demonstrates the critical value of full-text search
  • Provides a practical tool for scientific citation quality control

Rating

  • Novelty: ⭐⭐⭐⭐ Full-text search and iterative multi-citation recommendation are practical innovations
  • Experimental Thoroughness: ⭐⭐⭐⭐ Cross-domain evaluation + human annotation + multi-model comparison
  • Writing Quality: ⭐⭐⭐⭐ Clear problem definition, well-designed experiments