1 / 33

Differentially Private Recommendation Systems

18739A: Foundations of Security and Privacy. Differentially Private Recommendation Systems. Jeremiah Blocki Fall 2009. Netflix $1,000,000 Prize Competition. Queries: On a scale of 1 to 5 how would John rate “The Notebook” if he watched it?. Netflix Prize Competition.

ovid
Télécharger la présentation

Differentially Private Recommendation Systems

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. 18739A: Foundations of Security and Privacy Differentially Private Recommendation Systems Jeremiah Blocki Fall 2009

  2. Netflix $1,000,000 Prize Competition Queries: On a scale of 1 to 5 how would John rate “The Notebook” if he watched it?

  3. Netflix Prize Competition • Note: N x M table is very sparse (M = 17,770 movies, N = 500,000 users) • To Protect Privacy: • Each user was randomly assigned to a globally unique ID • Only 1/10 of the ratings were published • The ratings that were published were perturbed a little bit

  4. Outline • Breaking the anonymity of the Netflix prize dataset • Recap: Differential Privacy and define Approximate Differential Privacy • Prediction Algorithms • Privacy Preserving Prediction Algorithms • Remaining Issues

  5. Netflix FAQ: Privacy Source: How to Break Anonymity of the Netflix Prize Dataset (Narayanan and Shmatikov)

  6. De-Anonymizing the Netflix Dataset Source: How to Break Anonymity of the Netflix Prize Dataset (Narayanan and Shmatikov)

  7. De-Anonymizing the Netflix Dataset • Does anyone really care about the privacy of movie ratings? • Video Privacy Protection Act of 1988 exists, so some people do. • In some cases knowing movie ratings allows you to infer an individual’s political affiliation, religion, or even sexual orientation. Source: How to Break Anonymity of the Netflix Prize Dataset (Narayanan and Shmatikov)

  8. Privacy in Recommender Systems • Netflix might base its recommendation to me on both: • My own rating history • The rating history of other users • Maybe it isn’t such a big deal if the recommendations Netflix makes to me leak information about my own history, but they should preserve the privacy of other users • It would be possible to learn about someone else’s ratings history • If we have auxiliary information • Active attacks are a source auxiliary information Calandrino, et al. Don’t review that book: Privacy risks of collaborative filtering, 2009.

  9. Recall Differential Privacy [Dwork et al 2006] Randomized sanitization function κ has ε-differential privacy if for all data sets D1 and D2 differing by at most one element and all subsets S of the range of κ, Pr[κ(D1) S ] ≤ eε Pr[κ(D2) S ]

  10. Tell me f(D) Achieving Differential Privacy • How much noise should be added? • Intuition: f(D) can be released accurately when f is insensitive to individual entries x1, … xn (the more sensitive f is, higher the noise added) User Database D x1 … xn f(D)+noise

  11. Sensitivity of a function • Examples: count  1, histogram  1 • Note: f does not depend on the database

  12. Sensitivity with Laplace Noise

  13. Approximate Differential Privacy Randomized sanitization function κ has (ε, δ)-differential privacyif for all data sets D1 and D2 differing by at most one element and all subsets S of the range of κ, Pr[κ(D1)∈ S] ≤ eε Pr[κ(D2) S ]+ δ Source: Dwork et al 2006

  14. Approximate Differential Privacy • One interpretation of approximate differential privacy is that the outcomes of the computation κ are unlikely to provide much more information than ε-differential privacy, but it is possible. • Key Difference: Approximate Differential Privacy does NOT require that: Range(κ(D1)) = Range(κ(D2)) • The privacy guarantees made by (ε,δ)-differential privacy are not as strong as ε-differential privacy, but less noise is required to achieve (ε,δ)-differential privacy. Source: Dwork et al 2006

  15. Achieving Approximate Differential Privacy • Key Differences: • Use of the L2 instead of L1 norm to define the sensitivity of Δf • Use of Gaussian Noise instead of LaplacianNoise Source: Differentially Private Recommender Systems(McSherry and Mironov)

  16. Differential Privacy for Netflix Queries • What level of granularity to consider? What does it mean for databases D1and D2todiffer on at most one element? • One user (column) is present in D1 but not in D2 • One rating (cell) is present in D1 but not in D2 • Issue 1: Given a query “how would user i rate movie j?” Consider: K(D-u[i]) - how can it possibly be accurate? • Issue 2: If the definition of differing in at most one element is taken over cells, then what privacy guarantees are made for a user with many data points?

  17. Netflix Predictions – High Level • Q(i,j) – “How would user i rate movie j?” • Predicted rating may typically depend on • Global average rating over all movies and all users • Average movie rating of user i • Average rating of movie j • Ratings user i gave to similar movies • Ratings similar users gave to movie j • Intuition: Δf may be small for many of these queries

  18. Personal Rating Scale • For Alice a rating of 3 might mean the movie was really terrible. • For Bob the same rating might mean that the movie was excellent. • How do we tell the difference?

  19. How do we tell if two users are similar? • Pearson’s Correlation is one metric for similarity of users i and j • Consider all movies rated by both users • Negative value whenever i likes a movie that j dislikes • Positive value wheveri and j agreee • We can use similar metrics to measure the similarity between two movies.

  20. Netflix Predictions Example • Collaborative Filtering • Find the k-nearest neighbors of user i who have rated movie j by Pearson’s Correlation: • Predicted Rating similarity of users i and j k most similar users

  21. Netflix Prediction Sensitivity Example • Pretend the query Q(i,j) included user i’s rating history • At most one of the neighbors ratings changes, and the range of ratings is 4. The L1 sensitivity of the prediction is: p = 4/k

  22. Similarity of Two Movies • Let U be the set of all users who have rated both movies i and j then

  23. K-Nearest Users or K-Nearest Movies? Either way, after some pre-computation, we need to be able to find the k-nearest users/movies quickly!

  24. Covariance Matrix

  25. What do we need to make predictions? For a large class of prediction algorithms it suffices to have: • Gavg • Mavg – for each movie • Average Movie Rating for each user • Movie-Movie Covariance Matrix (COV)

  26. Differentially Private Recommender Systems (High Level) To respect approximate differential privacy publish • Gavg + NOISE • Mavg + NOISE • COV + NOISE • Don’t publish average ratings for users • Gavg, Mavg are very small so they can be published with little noise • COV requires more care. Source: Differentially Private Recommender Systems(McSherry and Mironov)

  27. Covariance Matrix Tricks • Trick 1: Center and clamp all ratings around averages. If we use clamped ratings then we reduce the sensitivity of our function.

  28. Covariance Matrix Tricks • Trick 2: Carefully weight the contribution of each user to reduce the sensitivity of the function. Users who have rated more movies are assigned lower weight.

  29. Publishing the Covariance Matrix • Theorem*: Pick then * Some details about stabilized averages are being swept under the rug

  30. Results Source: Differentially Private Recommender Systems(McSherry and Mironov)

  31. Note About Results • Granularity: One rating present in D1 but not in D2 • Accuracy is much lower when one user is present in D1 but not in D2 • Intuition: Given query Q(i,j) the database D-u[i] gives us no history about user i. • Approximate Differential Privacy • Gaussian Noise added according to L2 Sensitivity • Clamped Ratings (B =1) to further reduce noise

  32. Questions • Could the results be improved if the analysis was tailored to more specific prediction algorithms? • What if new movies are added? Can the movie-movie covariance matrix be changed without comprising privacy guarantees? • How strong are the privacy guarantees given that there are potentially many data points clustered around a single user?

More Related