Skip to content
Longterm Wiki
Back

MIRI Blog (Machine Intelligence Research Institute)

web

Credibility Rating

3/5
Good(3)

Good quality. Reputable source with community review or editorial standards, but less rigorous than peer-reviewed venues.

Rating inherited from publication venue: MIRI

MIRI is one of the oldest dedicated AI safety research organizations; their blog is a key primary source for foundational alignment research and organizational perspectives on long-term AI risk.

Metadata

Importance: 72/100blog posthomepage

Summary

The official blog of the Machine Intelligence Research Institute (MIRI), covering technical AI safety research including agent foundations, decision theory, logical uncertainty, and alignment. Posts range from research updates and technical results to broader reflections on the AI safety landscape and existential risk from advanced AI systems.

Key Points

  • Primary outlet for MIRI's technical research updates on agent foundations, decision theory, and logical uncertainty
  • Covers foundational alignment problems such as corrigibility, embedded agency, and value alignment
  • Includes strategic and philosophical posts about the risks posed by advanced AI and MIRI's research agenda
  • Features contributions from researchers like Eliezer Yudkowsky, Nate Soares, and Paul Christiano collaborators
  • Serves as a living record of MIRI's evolving research priorities and organizational thinking over many years

Cited by 1 page

PageTypeQuality
Machine Intelligence Research Institute (MIRI)Organization50.0

Cached Content Preview

HTTP 200Fetched Apr 9, 20263 KB
Blog - Machine Intelligence Research Institute 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

 

 
 
 
 
 
 

 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

 
 

 
 

 
 
 
 
 
 
 
 
 

 
 
 
 
 Skip to content 

 
 
 

 
 
 
 
 
 
 
 
 
 
 
 
 
 MIRI Updates

 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 Promising Signals on AI Governance from China 

 
 
 
 
 
 
 
 
 April 6, 2026 
 
 

 
 
 
 Joe Rogero 
 
 

 
 
 
 Uncategorized 
 
 

 
 
 
 
 
 View the official memo here. China has consistently signaled a willingness to engage on global AI governance since at least 2017. This memo compiles key statements from the Chinese government and prominent figures demonstrating their desire to coordinate on the... 
 
 
 
 
 
 
 
 
 Read more 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 The AI Doc : Your Questions Answered 

 
 
 
 
 
 
 
 
 March 27, 2026 
 
 

 
 
 
 Alana Horowitz Friedman, Joe Rogero, Rob Bensinger and Stefan Mitikj 
 
 

 
 
 
 News 
 
 

 
 
 
 
 
 So you’ve just seen The AI Doc: Or How I Became an Apocaloptimist, and you suddenly have questions, lots of them. The 104-minute documentary (currently in theaters) takes viewers on a fast-paced tour through the many dimensions of the AI... 
 
 
 
 
 
 
 
 
 Read more 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 MIRI Newsletter #125 

 
 
 
 
 
 
 
 
 March 19, 2026 
 
 

 
 
 
 Alana Horowitz Friedman and Rob Bensinger 
 
 

 
 
 
 Newsletters 
 
 

 
 
 
 
 
 The AI Doc: Buy tickets and spread the word! On Thursday, March 26th, a major new AI documentary is coming out: The AI Doc: Or How I Became an Apocaloptimist. Tickets are on sale now. The movie is excellent, and... 
 
 
 
 
 
 
 
 
 Read more 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 Summary: Mechanisms to Verify International Agreements about AI Development 

 
 
 
 
 
 
 
 
 March 18, 2026 
 
 

 
 
 
 Joe Rogero 
 
 

 
 
 
 Analysis , Papers 
 
 

 
 
 
 
 
 If world leaders agree to halt or limit AI development, they will need to verify that other nations are keeping their commitments. To this end, it helps to know where AI chips are, how they’re used, and what the AIs... 
 
 
 
 
 
 
 
 
 Read more 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 A Reliability Engineer Reviews Frontier AI Research 

 
 
 
 
 
 
 
 
 December 11, 2025 
 
 

 
 
 
 Joe Rogero 
 
 

 
 
 
 Analysis 
 
 

 
 
 
 
 
 This is part of the MIRI Single Author Series. Pieces in this series represent the beliefs and opinions of their named authors, and do not claim to speak for all of MIRI. Before the machine learning revolution kicked AI into... 
 
 
 
 
 
 
 
 
 Read more 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 MIRI Comms is hiring 

 
 
 
 
 
 
 
 
 December 10, 2025 
 
 

 
 
 
 Duncan Sabien 
 
 

 
 
 
 News 
 
 

 
 
 
 
 
 See details and apply. In the wake of the success of Nate and Eliezer’s book, If Anyone Builds It, Everyone Dies, we have an opportunity to push through a lot of doors that have cracked open, and roll a lot... 
 
 
 
 
 
 
 
 
 Read more 
 

... (truncated, 3 KB total)
Resource ID: 7f2ba8f23aeb7cd3 | Stable ID: sid_gW2S7eBcRw