Differential privacy
webCredibility Rating
High quality. Established institution or organization with editorial oversight and accountability.
Rating inherited from publication venue: Microsoft
Differential privacy is increasingly relevant to AI safety discussions around data governance, model training on sensitive data, and compliance with privacy regulations; this MSR page serves as an overview of the concept and related research.
Metadata
Summary
This Microsoft Research publication covers differential privacy, a mathematical framework that provides rigorous privacy guarantees when analyzing or publishing statistical information about datasets. It ensures that the inclusion or exclusion of any single individual's data has minimal impact on the output, protecting individual privacy while enabling aggregate analysis. The framework has become a foundational technique in privacy-preserving machine learning and data governance.
Key Points
- •Differential privacy provides a formal mathematical definition of privacy, quantified by a parameter epsilon (ε) that bounds information leakage about individuals.
- •Enables data analysis and machine learning on sensitive datasets while providing provable privacy guarantees rather than ad-hoc protections.
- •Widely adopted in AI/ML pipelines (e.g., federated learning, model training) to prevent models from memorizing or leaking private training data.
- •Relevant to AI governance and compliance frameworks requiring demonstrable privacy protections in automated decision systems.
- •Represents a key technical tool for balancing data utility with individual privacy rights in large-scale AI deployments.
Cited by 1 page
| Page | Type | Quality |
|---|---|---|
| AI-Driven Concentration of Power | Risk | 65.0 |
Cached Content Preview
Differential Privacy - Microsoft Research
Skip to main content
Research
Publications
Code & data
People
Microsoft Research blog
Artificial intelligence
Audio & acoustics
Computer vision
Graphics & multimedia
Human-computer interaction
Human language technologies
Search & information retrieval
Data platforms and analytics
Hardware & devices
Programming languages & software engineering
Quantum computing
Security, privacy & cryptography
Systems & networking
Algorithms
Mathematics
Ecology & environment
Economics
Medical, health & genomics
Social sciences
Technology for emerging markets
Academic programs
Events & academic conferences
Microsoft Research Forum
Behind the Tech podcast
Microsoft Research blog
Microsoft Research Forum
Microsoft Research podcast
About Microsoft Research
Careers & internships
People
Emeritus program
News & awards
Microsoft Research newsletter
Africa
AI for Science
AI Frontiers
Asia-Pacific
Cambridge
Health Futures
India
Montreal
New England
New York City
Redmond
Applied Sciences
Mixed Reality & AI - Cambridge
Mixed Reality & AI - Zurich
Register: Research Forum
Microsoft Security
Azure
Dynamics 365
Microsoft 365
Microsoft Teams
Windows 365
Microsoft AI
Azure Space
Mixed reality
Microsoft HoloLens
Microsoft Viva
Quantum computing
Sustainability
Education
Automotive
Financial services
Government
Healthcare
Manufacturing
Retail
Find a partner
Become a partner
Partner Network
Microsoft Marketplace
Marketplace Rewards
Software development companies
Blog
Microsoft Advertising
Developer Center
Documentation
Events
Licensing
Microsoft Learn
Microsoft Research
View Sitemap
Differential Privacy
Cynthia Dwork
33rd International Colloquium on Automata, Languages and Programming, part II (ICALP 2006)
| July 2006
Published by Springer Verlag
Publication
Download BibTex
In 1977 Dalenius articulated a desideratum for statistical databases: nothing about an individual should be learnable from the database that cannot be learned without access to the database. We give a general impossibility result showing that a formalization of Dalenius’ goal along the lines of semantic security cannot be achieved. Contrary to intuition, a variant of the result threatens the privacy even of someone not in the database. This state of affairs suggests a new measure, differential privacy, which, intuitively, captures the increased risk to one’s privacy incurred by participating in a database. The techniques developed in a sequence of p
... (truncated, 5 KB total)d0dcb570edc50d34 | Stable ID: sid_BBAVnuMcyH