Skip to content
Longterm Wiki
Back

Credibility Rating

3/5
Good(3)

Good quality. Reputable source with community review or editorial standards, but less rigorous than peer-reviewed venues.

Rating inherited from publication venue: 80,000 Hours

This 80,000 Hours podcast episode is relevant to AI safety communities interested in broader catastrophic risk governance, as biosecurity shares structural challenges with AI safety including dual-use research dilemmas, international coordination failures, and the need for proactive policy before catastrophe occurs.

Metadata

Importance: 62/100podcast episodeeducational

Summary

Jaime Yassif, a biosecurity expert at the Nuclear Threat Initiative, discusses the risks of engineered pandemics and biological weapons, the dual-use dilemma in life sciences research, and strategies for improving global biosecurity governance. The conversation covers how advances in biotechnology could lower barriers to creating dangerous pathogens and what policy interventions could reduce catastrophic biological risks.

Key Points

  • Advances in synthetic biology and DNA synthesis are making it easier to engineer dangerous pathogens, lowering the barrier for both state and non-state actors.
  • The dual-use dilemma in bioscience means beneficial research can inadvertently provide knowledge or tools useful for bioweapons development.
  • Improving biosecurity requires coordination across international governance frameworks, screening of DNA synthesis orders, and enhanced laboratory biosafety standards.
  • Engineered pandemics are considered among the most severe global catastrophic risks, potentially on par with nuclear threats in terms of civilizational impact.
  • There is a significant talent and funding gap in biosecurity policy compared to the scale of the threat, making it a high-impact career area.

Cited by 1 page

Cached Content Preview

HTTP 200Fetched Apr 9, 202698 KB
Jaime Yassif on safeguarding bioscience to prevent catastrophic lab accidents and bioweapons development | 80,000 Hours Search for: Our new book, a ridiculously in-depth guide to a fulfilling career, is out May 2026. Preorder now 

 On this page:

 Introduction 
 1 Highlights 
 2 Articles, books, and other media discussed in the show 
 3 Transcript 3.1 Rob's intro [00:00:00] 
 3.2 The interview begins [00:02:32] 
 3.3 Categories of global catastrophic biological risks [00:05:24] 
 3.4 Disagreements with the effective altruism community [00:07:39] 
 3.5 Stopping the first person from getting infected [00:11:51] 
 3.6 Shaping intent [00:15:51] 
 3.7 Verification and the Biological Weapons Convention [00:25:34] 
 3.8 Attribution [00:37:19] 
 3.9 How to actually implement a new idea [00:50:58] 
 3.10 COVID-19: natural pandemic or lab leak? [00:53:35] 
 3.11 How much can we rely on traditional law enforcement to detect terrorists? [00:58:24] 
 3.12 Constraining capabilities [01:01:27] 
 3.13 The funding landscape [01:07:00] 
 3.14 Oversight committees [01:14:24] 
 3.15 Just winning the argument [01:20:20] 
 3.16 NTI's vision [01:27:43] 
 3.17 Suppliers of goods and services [01:33:27] 
 3.18 Publishers [01:39:45] 
 3.19 Biggest weaknesses of NTI platform [01:42:33] 
 3.20 Careers [01:48:35] 
 3.21 How people outside of the US can best contribute [01:54:14] 
 3.22 Academia vs think tanks vs nonprofits vs government [01:59:25] 
 3.23 International cooperation [02:05:44] 
 3.24 Best things about living in the US, UK, China, and Israel [02:11:19] 
 3.25 Rob's outro [02:14:58] 
 
 4 Learn more 
 5 Related episodes 
 Read transcript See all episodes 
 
 
 
 

 

 We should not have a single point of failure.

 If this is really part of a shared global effort to safeguard the future of humanity, we need to intervene at every point possible — prevention, detection, and response — to reduce the risk as close to zero as we can that we would face something catastrophic from a biological release in the future.

 — Jaime Yassif

 If a rich country were really committed to pursuing an active biological weapons program, there’s not much we could do to stop them. With enough money and persistence, they’d be able to buy equipment, and hire people to carry out the work.

 But what we can do is intervene before they make that decision.

 Today’s guest, Jaime Yassif — Senior Fellow for global biological policy and programs at the Nuclear Threat Initiative (NTI) — thinks that stopping states from wanting to pursue dangerous bioscience in the first place is one of our key lines of defence against global catastrophic biological risks (GCBRs).

 It helps to understand why countries might consider developing biological weapons. Jaime says there are three main possible reasons:

 Fear of what their adversary might be up to
 Belief that they could gain a tactical or strategic advantage, with limited risk of getting caught
 Belief that even if they 

... (truncated, 98 KB total)
Resource ID: f30426a3f361f0a3 | Stable ID: sid_aD0olBn8SA