Differential privacy
webCredibility Rating
High quality. Established institution or organization with editorial oversight and accountability.
Rating inherited from publication venue: Microsoft
Differential privacy is increasingly relevant to AI safety discussions around data governance, model training on sensitive data, and compliance with privacy regulations; this MSR page serves as an overview of the concept and related research.
Metadata
Summary
This Microsoft Research publication covers differential privacy, a mathematical framework that provides rigorous privacy guarantees when analyzing or publishing statistical information about datasets. It ensures that the inclusion or exclusion of any single individual's data has minimal impact on the output, protecting individual privacy while enabling aggregate analysis. The framework has become a foundational technique in privacy-preserving machine learning and data governance.
Key Points
- •Differential privacy provides a formal mathematical definition of privacy, quantified by a parameter epsilon (ε) that bounds information leakage about individuals.
- •Enables data analysis and machine learning on sensitive datasets while providing provable privacy guarantees rather than ad-hoc protections.
- •Widely adopted in AI/ML pipelines (e.g., federated learning, model training) to prevent models from memorizing or leaking private training data.
- •Relevant to AI governance and compliance frameworks requiring demonstrable privacy protections in automated decision systems.
- •Represents a key technical tool for balancing data utility with individual privacy rights in large-scale AI deployments.
Cited by 1 page
| Page | Type | Quality |
|---|---|---|
| AI-Driven Concentration of Power | Risk | 65.0 |
Cached Content Preview
 Skip to main content # Differential Privacy - Cynthia Dwork **_33rd International Colloquium on Automata, Languages and Programming, part II (ICALP 2006)_** \| July 2006 Published by Springer Verlag [Publication](http://dx.doi.org/10.1007/11787006_1) [Download BibTex](https://www.microsoft.com/en-us/research/publication/differential-privacy/bibtex/) In 1977 Dalenius articulated a desideratum for statistical databases: nothing about an individual should be learnable from the database that cannot be learned without access to the database. We give a general impossibility result showing that a formalization of Dalenius’ goal along the lines of semantic security cannot be achieved. Contrary to intuition, a variant of the result threatens the privacy even of someone not in the database. This state of affairs suggests a new measure, differential privacy, which, intuitively, captures the increased risk to one’s privacy incurred by participating in a database. The techniques developed in a sequence of papers \[8, 13, 3\], culminating in those described in \[12\], can achieve any desired level of privacy under this measure. In many cases, extremely accurate information about the database can be provided while simultaneously ensuring very high levels of privacy. Opens in a new tab Follow us: - [Follow on X](https://x.com/intent/follow?original_referrer=https%3A%2F%2Fwww.microsoft.com%2Fen-us%2Fresearch%2Fpublication%2Fdifferential-privacy%2F&screen_name=MSFTResearch) - [Like on Facebook](https://www.facebook.com/microsoftresearch/) - [Follow on LinkedIn](https://www.linkedin.com/showcase/microsoftresearch/) - [Subscribe on Youtube](https://www.youtube.com/user/MicrosoftResearch) - [Follow on Instagram](https://www.instagram.com/msft_research/) - [Subscribe to our RSS feed](https://www.microsoft.com/en-us/research/feed/) Share this page: - [Share on X](https://x.com/intent/tweet?text=Differential%20Privacy&url=https%3A%2F%2Fwww.microsoft.com%2Fen-us%2Fresearch%2Fpublication%2Fdifferential-privacy%2F) - [Share on Facebook](https://www.facebook.com/sharer/sharer.php?u=https%3A%2F%2Fwww.microsoft.com%2Fen-us%2Fresearch%2Fpublication%2Fdifferential-privacy%2F) - [Share on LinkedIn](https://www.linkedin.com/shareArticle?mini=true&url=https%3A%2F%2Fwww.microsoft.com%2Fen-us%2Fresearch%2Fpublication%2Fdifferential-privacy%2F&title=Differential%20Privacy&summary=Differential%20Privacy&source=Microsoft%20Research) - [Share on Reddit](http://www.reddit.com/submit?title=Differential%20Privacy&url=https%3A%2F%2Fwww.microsoft.com%2Fen-us%2Fresearch%2Fpublication%2Fdifferential-privacy%2F) Notifications
d0dcb570edc50d34 | Stable ID: NGYwOTc5MT