DeepMind Safety Research Publications
webCredibility Rating
High quality. Established institution or organization with editorial oversight and accountability.
Rating inherited from publication venue: Google DeepMind
This is DeepMind's official publications page filtered by 'safety', serving as a regularly updated index of their safety-relevant research output; useful for tracking the lab's priorities and finding primary sources on specific safety topics.
Metadata
Summary
A curated index of DeepMind/Google DeepMind research publications filtered by the 'safety' tag, covering 240 papers spanning topics such as AI consciousness, existential safety, human-AI alignment, AI personhood, and technical safety research. The listing spans multiple years and reflects the breadth of safety-related work coming out of one of the world's leading AI labs.
Key Points
- •240 publications tagged as safety-related from DeepMind, spanning 2025-2026 and earlier years
- •Includes foundational and exploratory topics: existential safety, AI consciousness, human-AI alignment, and AI personhood
- •'Imitation Learning is Probably Existentially Safe' and 'The Abstraction Fallacy' suggest engagement with core AI safety debates
- •Covers both technical safety research and sociotechnical/philosophical dimensions of AI safety
- •Useful as a living index to track DeepMind's evolving safety research agenda over time
Cited by 1 page
| Page | Type | Quality |
|---|---|---|
| AI Safety Research Value Model | Analysis | 60.0 |
Cached Content Preview
Publications — Google DeepMind Skip to main content Publications Explore a selection of our recent research on some of the most complex and interesting challenges in AI. 240 publications 23 April 2026 Dynamic Reflections: Probing Video Representations with Text Alignment 10 March 2026 The Abstraction Fallacy: Why AI Can Simulate But Not Instantiate Consciousness 15 February 2026 Simplicity and Complexity in Combinatorial Optimization 5 February 2026 Hybrid neural–cognitive models reveal how memory shapes human reward learning 9 January 2026 TRecViT: A Recurrent Video Transformer 21 November 2025 Imitation Learning is Probably Existentially Safe 4 November 2025 To Mask or to Mirror: Human-AI Alignment in Collective Reasoning 30 October 2025 A Pragmatic View of AI Personhood 29 September 2025 AI-Generated Video Detection via Perceptual Straightening 24 September 2025 Video models are zero-shot learners and reasoners 24 September 2025 EmbeddingGemma: Powerful and Lightweight Text Representations 4 September 2025 Improving cosmological reach of LIGO usingDeep Loop Shaping 3 September 2025 RoboBallet: Planning for Multi-Robot Reaching with Graph Neural Networks and Reinforcement Learning 8 August 2025 Properties of Algorithmic Information Distance 1 August 2025 Visual Intention Grounding for Egocentric Assistants 16 July 2025 Dialogues Between Technologists and the Art Worlds 13 July 2025 Large Language Models as Rankers, Judges, and Assistants: A Perspective on the Potential Over-Reliance on LLMs in IR 13 July 2025 SLIM: ONE-SHOT QUANTIZED SPARSE PLUS LOW-RANK APPROXIMATION OF LLMS 13 July 2025 Long-Form Speech Generation with Spoken Language Models 1 July 2025 Rethinking Example Selection in the Era of Million-Token Models 26 June 2025 Performance Prediction for Large Systems via Text-to-Text Regression 23 June 2025 LIA: Cost-efficient LLM Inference Acceleration with Intel Advanced Matrix Extensions and CXL 20 June 2025 AuPair: Golden Example Pairs for Code Repair 1 June 2025 Bridging Algorithmic Information Theory and Machine Learning, Part II: Clustering, Density Estimation, Kolmogorov Complexity-Based Kernels, and Kernel Learning in Unsupervised Learning 1 May 2025 Proactive Agents for Multi-Turn Text-to-Image Generation Under Uncertainty 29 April 2025 Prompting with Phonemes: Enhancing LLM Multilinguality for non-Latin Scripts 28 April 2025 Flow-Lenia: Emergent evolutionary dynamics in mass conservative continuous cellular automata 26 April 2025 Relaxed Recursive Transformers: Effective Parameter Sharing with Layer-wise LoRA 26 April 2025 Generative Ghosts: Anticipating Benefits and Risks of AI Afterlives 26 April 2025 Toward Understanding In-context vs. In-weight Learning 1 Page 2 Page 3 … Page 8
70b4461a02951e08 | Stable ID: sid_eGHN9DLHCc