Longterm Wiki

Reducing Hallucinations in AI-Generated Wiki Content

reducing-hallucinationsapproachPath: /knowledge-base/responses/reducing-hallucinations/
E814Entity ID (EID)
← Back to page0 backlinksQuality: 68Updated: 2026-03-13
Page Recorddatabase.json — merged from MDX frontmatter + Entity YAML + computed metrics at build time
{
  "id": "reducing-hallucinations",
  "numericId": "E814",
  "path": "/knowledge-base/responses/reducing-hallucinations/",
  "filePath": "knowledge-base/responses/reducing-hallucinations.mdx",
  "title": "Reducing Hallucinations in AI-Generated Wiki Content",
  "quality": 68,
  "readerImportance": 55,
  "researchImportance": null,
  "tacticalValue": null,
  "contentFormat": "article",
  "tractability": null,
  "neglectedness": null,
  "uncertainty": null,
  "causalLevel": null,
  "lastUpdated": "2026-03-13",
  "dateCreated": "2026-02-17",
  "llmSummary": "This comprehensive technical guide documents methods to reduce AI hallucinations in wiki content from 3-27% to 0-6% through RAG, verification techniques, and human oversight, though notes complete elimination remains impossible. The article provides extensive quantified evidence (40+ citations) showing that while techniques like RAG with quality sources can dramatically reduce errors, fundamental architectural limitations mean hallucinations persist even in advanced systems.",
  "description": "Technical and procedural strategies to ground AI-generated content in verified information and reduce factual errors in wiki articles",
  "ratings": {
    "novelty": 4,
    "rigor": 7,
    "actionability": 8,
    "completeness": 8
  },
  "category": "responses",
  "subcategory": "alignment-training",
  "clusters": [
    "ai-safety"
  ],
  "metrics": {
    "wordCount": 4151,
    "tableCount": 3,
    "diagramCount": 0,
    "internalLinks": 2,
    "externalLinks": 3,
    "footnoteCount": 0,
    "bulletRatio": 0.27,
    "sectionCount": 54,
    "hasOverview": true,
    "structuralScore": 12
  },
  "suggestedQuality": 80,
  "updateFrequency": null,
  "evergreen": false,
  "wordCount": 4151,
  "unconvertedLinks": [
    {
      "text": "Hallucination (artificial intelligence)",
      "url": "https://en.wikipedia.org/wiki/Hallucination_(artificial_intelligence",
      "resourceId": "e8702eeda428b13b",
      "resourceTitle": "Wikipedia: Hallucination (artificial intelligence)"
    },
    {
      "text": "arXiv:2401.01313",
      "url": "https://arxiv.org/abs/2401.01313",
      "resourceId": "5ecee4eb7e22949a",
      "resourceTitle": "arXiv: Survey of Hallucination Mitigation"
    }
  ],
  "unconvertedLinkCount": 2,
  "convertedLinkCount": 0,
  "backlinkCount": 0,
  "citationHealth": {
    "total": 69,
    "withQuotes": 57,
    "verified": 57,
    "accuracyChecked": 57,
    "accurate": 43,
    "inaccurate": 1,
    "avgScore": 0.9854037970827337
  },
  "hallucinationRisk": {
    "level": "low",
    "score": 30,
    "factors": [
      "no-citations",
      "high-rigor",
      "conceptual-content"
    ]
  },
  "entityType": "approach",
  "redundancy": {
    "maxSimilarity": 16,
    "similarPages": [
      {
        "id": "language-models",
        "title": "Large Language Models",
        "path": "/knowledge-base/capabilities/language-models/",
        "similarity": 16
      },
      {
        "id": "reasoning",
        "title": "Reasoning and Planning",
        "path": "/knowledge-base/capabilities/reasoning/",
        "similarity": 16
      },
      {
        "id": "ai-forecasting",
        "title": "AI-Augmented Forecasting",
        "path": "/knowledge-base/responses/ai-forecasting/",
        "similarity": 16
      },
      {
        "id": "scalable-oversight",
        "title": "Scalable Oversight",
        "path": "/knowledge-base/responses/scalable-oversight/",
        "similarity": 16
      },
      {
        "id": "agentic-ai",
        "title": "Agentic AI",
        "path": "/knowledge-base/capabilities/agentic-ai/",
        "similarity": 15
      }
    ]
  },
  "changeHistory": [
    {
      "date": "2026-02-18",
      "branch": "claude/resolve-issue-251-XhJkg",
      "title": "Remove legacy pageTemplate frontmatter",
      "summary": "Removed the legacy `pageTemplate` frontmatter field from 15 MDX files. This field was carried over from the Astro/Starlight era and is not used by the Next.js application.",
      "model": "opus-4-6",
      "duration": "~10min"
    },
    {
      "date": "2026-02-17",
      "branch": "claude/reduce-hallucinations-2XdJp",
      "title": "Add hallucination reduction research page",
      "summary": "Created a new wiki page on reducing hallucinations in AI-generated wiki content using the Crux content pipeline. The page covers RAG, WikiChat, prompt engineering, verification techniques, fine-tuning, human-in-the-loop systems, and their limitations, with 71 citations from Perplexity research.",
      "pr": 189
    }
  ],
  "coverage": {
    "passing": 6,
    "total": 13,
    "targets": {
      "tables": 17,
      "diagrams": 2,
      "internalLinks": 33,
      "externalLinks": 21,
      "footnotes": 12,
      "references": 12
    },
    "actuals": {
      "tables": 3,
      "diagrams": 0,
      "internalLinks": 2,
      "externalLinks": 3,
      "footnotes": 0,
      "references": 8,
      "quotesWithQuotes": 57,
      "quotesTotal": 69,
      "accuracyChecked": 57,
      "accuracyTotal": 69
    },
    "items": {
      "llmSummary": "green",
      "schedule": "red",
      "entity": "green",
      "editHistory": "green",
      "overview": "green",
      "tables": "amber",
      "diagrams": "red",
      "internalLinks": "amber",
      "externalLinks": "amber",
      "footnotes": "red",
      "references": "amber",
      "quotes": "green",
      "accuracy": "green"
    },
    "editHistoryCount": 2,
    "ratingsString": "N:4 R:7 A:8 C:8"
  },
  "readerRank": 268,
  "recommendedScore": 185.33
}
External Links

No external links

Backlinks (0)

No backlinks

Longterm Wiki