Intra-Coalition Engagement Strategies
intra-coalition-engagement-strategiesconceptPath: /knowledge-base/cruxes/intra-coalition-engagement-strategies/
E2066Entity ID (EID)
Page Recorddatabase.json — merged from MDX frontmatter + Entity YAML + computed metrics at build time
{
"id": "intra-coalition-engagement-strategies",
"wikiId": "E2066",
"path": "/knowledge-base/cruxes/intra-coalition-engagement-strategies/",
"filePath": "knowledge-base/cruxes/intra-coalition-engagement-strategies.mdx",
"title": "Intra-Coalition Engagement Strategies",
"quality": null,
"readerImportance": null,
"researchImportance": null,
"tacticalValue": null,
"contentFormat": "article",
"causalLevel": null,
"lastUpdated": "2026-03-25",
"dateCreated": "2026-03-25",
"summary": "This article covers general coalition engagement theory (communication, trust, role clarity, consensus) drawn from community health and political science literature, with a thin section attempting to apply it to AI safety organizations; the connection to AI risk is tenuous and the sourcing is entirely opaque with no URLs provided. The general content is competent but largely standard coalition management literature with minimal novel insight for AI safety practitioners.",
"description": "Systematic approaches for maintaining active participation, cooperation, and commitment among members within a coalition working toward shared goals.",
"ratings": null,
"category": "cruxes",
"subcategory": null,
"clusters": [
"ai-safety"
],
"metrics": {
"wordCount": 2236,
"tableCount": 1,
"diagramCount": 0,
"internalLinks": 5,
"externalLinks": 0,
"footnoteCount": 11,
"bulletRatio": 0.05,
"sectionCount": 21,
"hasOverview": true,
"structuralScore": 13
},
"suggestedQuality": 87,
"updateFrequency": null,
"evergreen": true,
"wordCount": 2236,
"unconvertedLinks": [],
"unconvertedLinkCount": 0,
"convertedLinkCount": 0,
"backlinkCount": 0,
"hallucinationRisk": {
"level": "medium",
"score": 35,
"factors": [
"few-external-sources",
"moderately-cited",
"conceptual-content",
"mostly-unsourced-footnotes"
],
"integrityIssues": [
"mostly-unsourced-footnotes"
]
},
"entityType": "concept",
"redundancy": {
"maxSimilarity": 16,
"similarPages": [
{
"id": "ai-non-extremization-coordination",
"title": "AI Non-Extremization Coordination",
"path": "/knowledge-base/responses/ai-non-extremization-coordination/",
"similarity": 16
},
{
"id": "us-aisi",
"title": "US AI Safety Institute (now CAISI)",
"path": "/knowledge-base/organizations/us-aisi/",
"similarity": 15
},
{
"id": "ai-safety-institutes",
"title": "AI Safety Institutes",
"path": "/knowledge-base/responses/ai-safety-institutes/",
"similarity": 15
},
{
"id": "intervention-evaluation-for-political-stability",
"title": "Intervention Evaluation for Political Stability",
"path": "/knowledge-base/responses/intervention-evaluation-for-political-stability/",
"similarity": 15
},
{
"id": "nist-ai-rmf",
"title": "NIST AI Risk Management Framework",
"path": "/knowledge-base/responses/nist-ai-rmf/",
"similarity": 15
}
]
},
"coverage": {
"passing": 4,
"total": 13,
"targets": {
"tables": 9,
"diagrams": 1,
"internalLinks": 18,
"externalLinks": 11,
"footnotes": 7,
"references": 7
},
"actuals": {
"tables": 1,
"diagrams": 0,
"internalLinks": 5,
"externalLinks": 0,
"footnotes": 11,
"references": 0,
"quotesWithQuotes": 0,
"quotesTotal": 0,
"accuracyChecked": 0,
"accuracyTotal": 0
},
"items": {
"summary": "green",
"schedule": "red",
"entity": "green",
"editHistory": "red",
"overview": "green",
"tables": "amber",
"diagrams": "red",
"internalLinks": "amber",
"externalLinks": "red",
"footnotes": "green",
"references": "red",
"quotes": "red",
"accuracy": "red"
}
},
"recommendedScore": 21.82
}External Links
No external links
Backlinks (0)
No backlinks