Back
AI-Plans.com – Crowdsourced Peer Review Platform for AI Alignment Research
webmanifund.org·manifund.org/projects/ai-planscom-
AI-Plans.com is a crowdsourced peer-review platform for AI alignment research proposals, enabling structured critique and scoring of alignment plans to accelerate feedback loops in the AI safety community.
Metadata
Importance: 38/100tool pagetool
Summary
AI-Plans.com is a platform hosting 100+ AI alignment plans with a structured critique and voting system to surface vulnerabilities and strengths in proposed alignment approaches. It uses a karmic voting system to weight feedback by trusted users and hosts bi-monthly Critique-a-Thon events. The project seeks funding via Manifund to pay team members and fund prize pools for critique events.
Key Points
- •Hosts 100+ alignment plans with 150+ critiques, scored by strength vs. vulnerability critiques using a karmic voting system.
- •Bi-monthly Critique-a-Thon events incentivize community participation with leaderboards and badges.
- •Lowers barrier to contribution: critiquing is easier than plan-writing, voting is easier than critiquing.
- •Has produced detailed critiques of high-profile papers including OpenAI's Superalignment paper and DeepMind papers.
- •Seeking $20,000 in funding for team compensation and Critique-a-Thon prize funds; currently has 6 developers.
Cached Content Preview
HTTP 200Fetched Apr 11, 20267 KB
Feb
MAR
Apr
05
2025
2026
2027
success
fail
About this capture
COLLECTED BY
Collection: Common Crawl
Web crawl data from Common Crawl.
TIMESTAMPS
The Wayback Machine - http://web.archive.org/web/20260305210351/https://manifund.org/projects/ai-planscom-
Manifund
Home
Login
About
People
Categories
Newsletter
Home
About
People
Categories
Login
Create
AI-Plans.com | Manifund
5
AI-Plans.com
Science & technology
Technical AI safety
AI governance
Kabir Kumar
Active
Grant
$5,370raised
$20,000funding goal
Donate
Sign in to donate
p]:prose-li:my-0 text-gray-900 prose-blockquote:text-gray-600 prose-a:font-light prose-blockquote:font-light font-light break-anywhere empty:prose-p:after:content-["\00a0"]">
Project summary
AI-Plans.com is a rapidly growing platform for feedback on AI Alignment research.
As of January 2024, there are 100+ alignment plans on the site, with 150+ Critiques.
We hold bi-monthly Critique-a-Thon events, for which participation has continued to increase.
It’s extremely useful for many reasons:
- Showcases just how many vulnerabilities there are in all the current alignment plans
- Drastically improves the feedback loop for AI Alignment researchers
- Makes it much easier to contribute to AI Safety research
- Provides credentials for anyone looking to get started in AI Safety (badges and position on leaderboard)
On the site, all alignment plans are scored and ranked from highest to lowest, with new plans always starting at the top. Users vote on the critiques rather than on the plans themselves. Plans are then scored by the sum of the scores of Strength Critiques minus the sum of the scores of Vulnerability Critiques.
We use a karmic voting system which gives more weight to votes cast by more trusted (i.e. more upvoted and less downvoted) users. Users are incentivized with a leaderboard and badges.
The author or poster of a plan can iterate on their plan by selecting critiques to address and creating a new version.
There are several new features coming in a rebuild, which is currently being worked on, including:
- Sub critiques
- Annotations without additional sign-up (Currently, annotating on a post requires signing in with a hypothes.is account)
We also run Critique-a-Thons of AI Alignment plans. See the results of the December Critique-a-Thon here: https://aiplans.substack.com/p/december-critique-a-thon-results
We were the first ones to release detailed Critiques of the recent Superalignment paper by OpenAI and the recent DeepMind papers
What are this project's goals and how will you achieve them?
AI-Plans.com aims to accelerate AI alignment research via a focused feedback loop of public peer review.
The site is designed to elicit high-quality feedback from an open community of alignment researchers and enthusiast
... (truncated, 7 KB total)Resource ID:
b22fc44d61e22753 | Stable ID: sid_VceuvmtReT