Differentially Private Counts with Additive Constraints
Date
2021
Authors
Advisors
Journal Title
Journal ISSN
Volume Title
Repository Usage Stats
views
downloads
Abstract
Differential privacy is a rigorous mathematical definition of privacy, which guarantees that data analysis should not be able to distinguish whether any individual is in or out of the dataset. To reduce disclosure risks, statistical agencies and other organizations can release noisy counts that satisfy differential privacy. In some contexts, the released counts need to satisfy additive constraints; for example, the released value of a total should equal the sum of the released values of its components. In this thesis, I present a simple post-processing procedure for satisfying such additive constraints. The basic idea is (i) compute approximate posterior modes or find posterior samples of the true counts given the noisy counts, (ii) construct a multinomial distribution with trial size equal to the posterior mode or the posterior draw of the total and probability vector equal to fractions derived from the posterior modes or the posterior draws of the components, and (iii) find and release a mode or samples of this multinomial distribution. I also present an approach for making Bayesian inferences about the true counts given these post-processed, differentially private counts. I illustrate these methods using simulations.
Type
Department
Description
Provenance
Subjects
Citation
Permalink
Citation
Wang, Ziang (2021). Differentially Private Counts with Additive Constraints. Master's thesis, Duke University. Retrieved from https://hdl.handle.net/10161/23320.
Collections
Except where otherwise noted, student scholarship that was shared on DukeSpace after 2009 is made available to the public under a Creative Commons Attribution / Non-commercial / No derivatives (CC-BY-NC-ND) license. All rights in student work shared on DukeSpace before 2009 remain with the author and/or their designee, whose permission may be required for reuse.