Skip to content

LongtermWiki Strategy Brainstorm

Status: Working document, not polished Purpose: Think through what could go wrong and what success looks like


What is LongtermWiki actually trying to be?

Loading diagram...

These aren’t mutually exclusive, but they have very different implications for what we build and how we measure success.


The fear: We build a comprehensive knowledge base that nobody uses because:

  • It’s not differentiated from LessWrong, Wikipedia, 80K, etc.
  • Maintenance burden exceeds our capacity
  • Information gets stale and trust erodes
  • No clear “job to be done” that it uniquely serves

Signs we’re heading here:

  • Low repeat visitors
  • People cite primary sources instead of our pages
  • “Oh yeah I’ve seen that site” but no behavior change
  • Content quality diverges wildly across pages

Failure Mode 2: “Insights That Aren’t”

Section titled “Failure Mode 2: “Insights That Aren’t””

The fear: We try to generate novel strategic insights but:

  • We’re not actually smarter than the existing field
  • “Insights” are obvious to experts, only novel to us
  • Analysis is shallow because we’re spread too thin
  • We mistake complexity for depth

Signs we’re heading here:

  • Experts are unimpressed or dismissive
  • Our “insights” don’t survive contact with counterarguments
  • We can’t point to decisions that changed because of our work
  • Internal feeling of “are we just rearranging deck chairs?”

Failure Mode 3: “Cathedral in the Desert”

Section titled “Failure Mode 3: “Cathedral in the Desert””

The fear: We build something beautiful and comprehensive that nobody asked for:

  • Solves a problem that isn’t actually a bottleneck
  • Users don’t have the “job to be done” we imagined
  • The people who need prioritization help aren’t reading documents
  • Decision-makers use networks and calls, not wikis

Signs we’re heading here:

  • We can’t name 10 specific people who would use this weekly
  • User research reveals different pain points than we assumed
  • “This is cool but I wouldn’t actually use it”
  • Building for an imagined user rather than real ones

The fear: Initial build is fine but ongoing maintenance is unsustainable:

  • Content rots faster than we can update it
  • Quality degrades as original authors leave
  • Scope creep makes everything shallow
  • Becomes a zombie project that’s “still around” but not useful

Signs we’re heading here:

  • Growing backlog of “needs review” pages
  • Key pages haven’t been touched in 6+ months
  • New content added but old content not maintained
  • Feeling of dread when thinking about updates

The fear: The crux-mapping / worldview-mapping approach is too unusual:

  • Users don’t understand how to use it
  • Requires too much buy-in to a novel framework
  • Experts don’t want to be “mapped” onto worldview archetypes
  • The structure imposes false precision on messy disagreements

Signs we’re heading here:

  • People engage with the wiki content, ignore the crux structure
  • Pushback on worldview categories (“that’s not what I believe”)
  • The novel features feel gimmicky
  • Simpler alternatives would serve users better

This is the crux. Different definitions lead to very different projects:

Test: Can we point to funding decisions, research directions, or career choices that changed because of LongtermWiki?

AspectImplication
WorkflowNeed to be embedded in actual decision-making
RelationshipsDirect relationships with funders/researchers
QualityQuality of analysis matters more than breadth
ScaleMight only need to serve a small number of users well

Concerns: Very high bar, attribution is hard, might be serving decisions that would have happened anyway.

Test: Do users report that they understand AI safety landscape better after using LongtermWiki?

AspectImplication
ValueEducational value is primary
CoverageBreadth and accessibility matter
CompetitionCompetes with AI Safety Fundamentals, 80K, etc.
SuccessPeople recommend it to newcomers

Concerns: Lots of competition, “understanding” doesn’t necessarily lead to better decisions, risk of being a stepping stone people quickly move past.

Test: Do people use LongtermWiki’s categories and cruxes when discussing AI safety strategy?

AspectImplication
ProductThe framework itself is the product
SuccessLongtermWiki vocabulary becomes common
FocusCrux definitions, worldview archetypes
ContentLess about content, more about structure

Concerns: Very hard to achieve, might impose bad structure on good discourse, requires significant field buy-in.

Test: Do we help people identify where they disagree and why?

AspectImplication
FocusDisagreement mapping is primary
RepresentationNeed to represent multiple perspectives fairly
ValueReducing “talking past each other”
FormatCould be more interactive/tool-like

Concerns: Might reify disagreements instead of resolving them, hard to represent views fairly, unclear who the user is.

Test: Do we produce novel strategic insights that weren’t obvious before?

AspectImplication
NatureWe’re doing original analysis
QualityQuality of thinking matters most
FormatMight look more like reports than wiki
Success”I hadn’t thought of it that way”

Concerns: Are we actually capable of this? Might be better done by existing researchers. High variance in outcomes.


Focus on one very specific use case and nail it.

Example: “LongtermWiki helps funders compare interventions under different worldviews”

  • 30-50 intervention pages, deeply analyzed
  • Clear worldview → priority tool
  • Explicit targeting of Open Phil, SFF, smaller funders
  • Success = funders actually reference it in decision memos

Pros: Clear focus, measurable success, defensible niche Cons: Small user base, high stakes per user, might not be what funders want

Comprehensive reference that’s clearly better than alternatives.

Example: “LongtermWiki is the best single source for ‘what is X in AI safety?’”

  • 200+ pages, all at consistent quality
  • Good SEO, clear navigation
  • First stop for researchers, journalists, newcomers
  • Success = high traffic, people link to us

Pros: Clear value prop, measurable, scales well Cons: Maintenance burden, commodity competition, doesn’t leverage our unique angle

Not neutral, not comprehensive — actively opinionated about what matters.

Example: “LongtermWiki is where you go to understand the strategic landscape according to [specific perspective]”

  • Explicitly represents a worldview or analytical lens
  • Quality of argument matters more than coverage
  • More like a think tank than a wiki
  • Success = “LongtermWiki’s take on X” becomes a reference point

Pros: Differentiated, lower maintenance, can be higher quality per page Cons: Alienates those who disagree, requires us to actually have good takes, might be seen as biased

Focus almost entirely on the crux-mapping innovation.

Example: “LongtermWiki maps the key uncertainties in AI safety and what would resolve them”

  • Minimal wiki content, max crux structure
  • Interactive disagreement exploration tools
  • Integrate with prediction markets, expert surveys
  • Success = cruxes get referenced, forecasts get made

Pros: Novel, potentially high-impact, differentiated Cons: Unproven value, might be too weird, hard to measure success

Regular, updating analysis of AI safety landscape.

Example: “LongtermWiki provides quarterly strategic assessments of the AI safety field”

  • More report-like than wiki-like
  • Regular update cadence with clear “what changed”
  • Track predictions, update estimates
  • Success = people read the quarterly updates, cite trends

Pros: Clear rhythm, natural freshness, can be event-driven Cons: High ongoing effort, journalism-like, competes with newsletters


Candidate UserTheir NeedOur Fit
Funders (Open Phil, SFF)Compare interventionsMaybe? Do they want this?
Researchers choosing topicsUnderstand landscapeProbably already do this
Newcomers to fieldGet orientedStrong competition exists
Journalists/policymakersQuick referenceMight be underserved
AI labs making safety decisions???Probably not us

How to resolve: Actually talk to potential users.

Uncertainty 2: Are we capable of generating insights?

Section titled “Uncertainty 2: Are we capable of generating insights?”

Honest assessment: Do we have the expertise to say things that aren’t obvious to field insiders?

  • If yes → lean into analysis and synthesis
  • If no → lean into curation and structure

How to resolve: Try it with a few topics and get expert feedback.

Uncertainty 3: Is the crux-mapping frame valuable?

Section titled “Uncertainty 3: Is the crux-mapping frame valuable?”

The LongtermWiki vision heavily features worldviews, cruxes, disagreement mapping. Is this actually useful or is it an intellectual hobby?

  • If useful → it’s our core differentiator
  • If not → we’re just a wiki with extra steps

How to resolve: Prototype the crux interface, test with users.

2 person-years builds it. What maintains it?

  • Ongoing funding?
  • Community contribution?
  • AI assistance?
  • Graceful archiving?

How to resolve: Plan for maintenance from day 1, or plan for finite lifespan.


Before going all-in, we could test key assumptions:

TestDurationWhat We Learn
User Interviews2 weeksWho wants this? What do they actually need?
Crux Prototype2 weeksDoes crux-mapping work in practice?
Page Quality Test2 weeksIs our content better than alternatives?
Insight Generation Test2 weeksCan we produce valuable novel insights?

Talk to 10-15 potential users:

  • Funders, researchers, policy people
  • “How do you currently make prioritization decisions?”
  • “What information do you wish you had?”
  • “Would you use X if it existed?”

Build a minimal crux-mapping interface for 3-5 cruxes:

  • Show to experts, get feedback
  • “Does this capture the disagreement?”
  • “Would you use this?”

Write 5 pages at our target quality level:

  • Show to potential users
  • “Is this useful? Better than alternatives?”
  • “What’s missing?”

Try to generate 3 novel strategic insights:

  • Write them up
  • Share with experts
  • “Is this valuable? Novel? Correct?”

  1. The wiki is necessary but not sufficient. We need good reference content, but that alone won’t differentiate us.

  2. The crux-mapping is our unique angle but it’s also the riskiest part. Need to validate it works.

  3. Narrow is safer than broad. Better to serve 50 users very well than 500 users poorly.

  4. We should pick a user and work backwards. Abstract “field building” goals are too vague.

  5. Maintenance is the hard part. Initial build is straightforward; sustainability is the real challenge.

  6. We might be solving a problem that isn’t actually a bottleneck. Need to validate that prioritization confusion is actually causing suboptimal resource allocation.


  1. Who are 10 specific people who would use this weekly? Can we name them?

  2. What’s the simplest version of LongtermWiki that would still be valuable?

  3. If we could only do one of (wiki, crux map, worldview tool), which?

  4. What would make us confident this is worth 2 person-years vs. not?

  5. Are there existing projects we should join/support rather than build?

  6. What’s the “failure mode” that would make us kill the project?

  7. How do we avoid this becoming a self-justifying project that continues because it exists?


  • Schedule user interviews with 5-10 potential users
  • Define “minimum viable LongtermWiki” that could be tested in 4 weeks
  • Identify 3 specific cruxes to prototype mapping
  • Honest assessment: do we have the expertise for Option 3 (Opinionated Synthesis)?
  • Research what happened to similar past projects ✓ (see Similar Projects Analysis)