Statistics – Machine Learning
Scientific paper
2009-01-10
Statistics
Machine Learning
14 pages
Scientific paper
This work studies formal utility and privacy guarantees for a simple multiplicative database transformation, where the data are compressed by a random linear or affine transformation, reducing the number of data records substantially, while preserving the number of original input variables. We provide an analysis framework inspired by a recent concept known as differential privacy (Dwork 06). Our goal is to show that, despite the general difficulty of achieving the differential privacy guarantee, it is possible to publish synthetic data that are useful for a number of common statistical learning applications. This includes high dimensional sparse regression (Zhou et al. 07), principal component analysis (PCA), and other statistical measures (Liu et al. 06) based on the covariance of the initial data.
Ligett Katrina
Wasserman Larry
Zhou Shuheng
No associations
LandOfFree
Differential Privacy with Compression does not yet have a rating. At this time, there are no reviews or comments for this scientific paper.
If you have personal experience with Differential Privacy with Compression, we encourage you to share that experience with our LandOfFree.com community. Your opinion is very important and Differential Privacy with Compression will most certainly appreciate the feedback.
Profile ID: LFWR-SCP-O-441051