Longterm Wiki
Back

Unsolved Problems in ML Safety

paper

Authors

Dan Hendrycks·Nicholas Carlini·John Schulman·Jacob Steinhardt

Credibility Rating

3/5
Good(3)

Good quality. Reputable source with community review or editorial standards, but less rigorous than peer-reviewed venues.

Rating inherited from publication venue: arXiv

Data Status

Not fetched

Abstract

Machine learning (ML) systems are rapidly increasing in size, are acquiring new capabilities, and are increasingly deployed in high-stakes settings. As with other powerful technologies, safety for ML should be a leading research priority. In response to emerging safety challenges in ML, such as those introduced by recent large-scale models, we provide a new roadmap for ML Safety and refine the technical problems that the field needs to address. We present four problems ready for research, namely withstanding hazards ("Robustness"), identifying hazards ("Monitoring"), reducing inherent model hazards ("Alignment"), and reducing systemic hazards ("Systemic Safety"). Throughout, we clarify each problem's motivation and provide concrete research directions.

Cited by 2 pages

PageTypeQuality
Center for AI SafetyOrganization42.0
Dan HendrycksPerson19.0
Resource ID: f94e705023d45765 | Stable ID: MzYxNmRmMj