Skip to content
Longterm Wiki
Back

Stanford HAI framework

web
crfm.stanford.edu·crfm.stanford.edu/open-fms/

Relevant to debates around open-source AI safety: whether releasing model weights openly increases risk from misuse or improves safety through broader scrutiny; produced by Stanford HAI's CRFM group, a leading academic institution in foundation model research.

Metadata

Importance: 58/100homepageanalysis

Summary

Stanford's Center for Research on Foundation Models (CRFM) presents a framework for evaluating and governing open foundation models, addressing the tradeoffs between openness and safety in large AI systems. It provides structured analysis of how open release policies affect safety, accountability, and beneficial use of foundation models.

Key Points

  • Examines the tradeoffs of open vs. closed release strategies for large foundation models and their implications for safety and misuse
  • Provides a governance framework for thinking about when and how foundation models should be released openly
  • Addresses accountability mechanisms for foundation model developers and downstream users
  • Considers dual-use risks and the challenge of preventing harmful applications of openly released models
  • Connects technical deployment decisions to broader policy and societal impact considerations

Cited by 1 page

PageTypeQuality
Open Source AI SafetyApproach62.0

Cached Content Preview

HTTP 200Fetched Apr 9, 202623 KB
On the Societal Impact of Open Foundation Models 

 
 
 
 
 
 

 
 
 
 
 

 

 

 

 

 
 
 
 
 
 On the Societal Impact of 
 Open Foundation Models 

 
 
 Analyzing the benefits and risks of foundation models with widely available weights 
 

 
 Paper (ICML 2024 Oral) 
 Blog 
 Policy brief 
 Authors 
 

 
 
 
 

 
 
 
 
 
 
 Context. One of the biggest tech policy debates today is about the future of AI,
 especially foundation models and generative AI. Should open AI models be restricted? This
 question is central to
 several policy efforts like the EU AI Act and the U.S. Executive Order on Safe, Secure, and
 Trustworthy AI.
 
 
 

 
 
 
 Status quo. Open foundation models, defined here as models with widely available weights,
 enable greater customization and deeper inspection.
 However, their downstream use cannot be monitored
 or moderated. As a result, risks relating to biosecurity, cybersecurity, disinformation, and
 non-consensual
 deepfakes have prompted pushback.
 
 
 

 
 
 
 Contributions. We analyze the benefits and risks of open foundation models. In
 particular, we present a framework to assess their marginal risk compared to closed
 models or
 existing technology. The framework helps explain why the marginal risk is low in some cases,
 clarifies disagreements in past studies by revealing the different assumptions about risk, and
 can help foster more
 constructive debate going forward.
 
 
 
 
 

 
 
 
 Key contributions

 
 Identifying distinctive properties. Foundation models released with widely available
 weights have
 distinctive properties that lead to both their benefits and risks. We outline five
 properties
 that inform our analysis of their societal impact: broader access, greater customizability, the
 ability for local inference and adaptability, an inability to rescind model weights once
 released, and an inability to monitor or moderate usage.

 Connecting properties to benefits. Open foundation models can distribute decision-making
 power, reduce market
 concentration,
 increase innovation, accelerate science, and enable transparency.
 We highlight considerations that may temper these benefits in practice (for example, model
 weights
 are
 sufficient for some forms of science, but access to training data is necessary for others and is
 not
 guaranteed by release of weights).

 Developing a risk assessment framework. We present a framework for conceptualizing the
 marginal
 risk of open foundation models:
 the extent to which these models increase societal risk by intentional misuse beyond
 closed foundation models or pre-existing technologies (such as web search on the internet).
 

 Re-assessing past studies. Surveying seven common misuse vectors described for open
 foundation
 models (such as disinformation,
 biosecurity, cybersecurity, non-consensual intimate imagery, scams), we find that past studies
 do
 not
 clearly assess the marginal risk in most cases.
 In particular, we encourage more grounded research on ch

... (truncated, 23 KB total)
Resource ID: a2e936ba01459a44 | Stable ID: sid_95f15qf4V4