auditing and ml privacy - stanford university · reading assignments (ml auditing) amodei, dario,...

47
Auditing and ML Privacy Jun 3, 2020 Dr. Wei Wei, Prof. James Landay CS 335: Fair, Accountable, and Transparent (FAccT) Deep Learning Stanford University

Upload: others

Post on 27-Jul-2020

3 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Auditing and ML Privacy - Stanford University · Reading Assignments (ML Auditing) Amodei, Dario, Chris Olah, Jacob Steinhardt, Paul Christiano, John Schulman, and Dan Mané. Concrete

Auditing and ML PrivacyJun 3, 2020

Dr. Wei Wei, Prof. James Landay

CS 335: Fair, Accountable, and Transparent (FAccT) Deep LearningStanford University

Page 2: Auditing and ML Privacy - Stanford University · Reading Assignments (ML Auditing) Amodei, Dario, Chris Olah, Jacob Steinhardt, Paul Christiano, John Schulman, and Dan Mané. Concrete

Outline● ML Auditing

○ Distill-and-Compare

● Privacy in ML○ Differential Privacy with Deep Learning○ Model Inversion Attack and Differential Privacy○ Local Differential Privacy○ Federated Learning

Page 3: Auditing and ML Privacy - Stanford University · Reading Assignments (ML Auditing) Amodei, Dario, Chris Olah, Jacob Steinhardt, Paul Christiano, John Schulman, and Dan Mané. Concrete

ML Auditing Using Model Distillation

Tan et al, 2018

Page 4: Auditing and ML Privacy - Stanford University · Reading Assignments (ML Auditing) Amodei, Dario, Chris Olah, Jacob Steinhardt, Paul Christiano, John Schulman, and Dan Mané. Concrete

General Additive Model

transformatione.g., logistic for classification

weights

Page 5: Auditing and ML Privacy - Stanford University · Reading Assignments (ML Auditing) Amodei, Dario, Chris Olah, Jacob Steinhardt, Paul Christiano, John Schulman, and Dan Mané. Concrete

Chicago Police “Strategic Subject”.● A risk score for individuals being victims or offenders in a shooting incident● 16 features

○ 8 reported being used by Chicago Police

Page 6: Auditing and ML Privacy - Stanford University · Reading Assignments (ML Auditing) Amodei, Dario, Chris Olah, Jacob Steinhardt, Paul Christiano, John Schulman, and Dan Mané. Concrete

Features Reported being Used

green - model being auditedred - mimic model

Page 7: Auditing and ML Privacy - Stanford University · Reading Assignments (ML Auditing) Amodei, Dario, Chris Olah, Jacob Steinhardt, Paul Christiano, John Schulman, and Dan Mané. Concrete

Features Reported Not Being Used

green - model being auditedred - mimic model

Page 8: Auditing and ML Privacy - Stanford University · Reading Assignments (ML Auditing) Amodei, Dario, Chris Olah, Jacob Steinhardt, Paul Christiano, John Schulman, and Dan Mané. Concrete

Auditing COMPAS

green - model being auditedred - mimic model

Page 9: Auditing and ML Privacy - Stanford University · Reading Assignments (ML Auditing) Amodei, Dario, Chris Olah, Jacob Steinhardt, Paul Christiano, John Schulman, and Dan Mané. Concrete

Auditing Lending Club

green - model being auditedred - mimic model

Page 10: Auditing and ML Privacy - Stanford University · Reading Assignments (ML Auditing) Amodei, Dario, Chris Olah, Jacob Steinhardt, Paul Christiano, John Schulman, and Dan Mané. Concrete

Outline● ML Auditing

○ Distill-and-Compare

● Privacy in ML○ Differential Privacy with Deep Learning○ Model Inversion Attack and Differential Privacy○ Local Differential Privacy○ Federated Learning

Page 11: Auditing and ML Privacy - Stanford University · Reading Assignments (ML Auditing) Amodei, Dario, Chris Olah, Jacob Steinhardt, Paul Christiano, John Schulman, and Dan Mané. Concrete

Privacy in ML

Training Data

Malicious Users

Deep Learning Models

Page 12: Auditing and ML Privacy - Stanford University · Reading Assignments (ML Auditing) Amodei, Dario, Chris Olah, Jacob Steinhardt, Paul Christiano, John Schulman, and Dan Mané. Concrete

Inferring Sensitive Features from ML Models

Fredrikson et al, 2014

Demographic InfoMedical HistoryGenetic Markers

Dose of Warfarin

Page 13: Auditing and ML Privacy - Stanford University · Reading Assignments (ML Auditing) Amodei, Dario, Chris Olah, Jacob Steinhardt, Paul Christiano, John Schulman, and Dan Mané. Concrete

Inferring Training Data from Facial Recognition Models

Fredrikson et al, 2015

Original Image Inferred Image

Page 14: Auditing and ML Privacy - Stanford University · Reading Assignments (ML Auditing) Amodei, Dario, Chris Olah, Jacob Steinhardt, Paul Christiano, John Schulman, and Dan Mané. Concrete

Centralized Setting

Compromised App

Healthy Apps DataModel

Page 15: Auditing and ML Privacy - Stanford University · Reading Assignments (ML Auditing) Amodei, Dario, Chris Olah, Jacob Steinhardt, Paul Christiano, John Schulman, and Dan Mané. Concrete

Distributed Setting

DataModel

Page 16: Auditing and ML Privacy - Stanford University · Reading Assignments (ML Auditing) Amodei, Dario, Chris Olah, Jacob Steinhardt, Paul Christiano, John Schulman, and Dan Mané. Concrete

Outline● ML Auditing

○ Distill-and-Compare

● Privacy in ML○ Differential Privacy with Deep Learning○ Model Inversion Attack and Differential Privacy○ Local Differential Privacy○ Federated Learning

Page 17: Auditing and ML Privacy - Stanford University · Reading Assignments (ML Auditing) Amodei, Dario, Chris Olah, Jacob Steinhardt, Paul Christiano, John Schulman, and Dan Mané. Concrete

Differential Privacy

Dataset D Dataset D'

+

M

M(D')

M

M(D)

Adjacent Inputs

Compare

Page 18: Auditing and ML Privacy - Stanford University · Reading Assignments (ML Auditing) Amodei, Dario, Chris Olah, Jacob Steinhardt, Paul Christiano, John Schulman, and Dan Mané. Concrete

Differential Privacy● A randomized mechanism satisfies (ε, δ) - differential privacy for adjacent

inputs d and d' if

Abadi et al, 2016

Dataset d Adjacent dataset d'

A small probability of failure

Amount of Information Leakage

Page 19: Auditing and ML Privacy - Stanford University · Reading Assignments (ML Auditing) Amodei, Dario, Chris Olah, Jacob Steinhardt, Paul Christiano, John Schulman, and Dan Mané. Concrete

Differential Privacy with Deep Learning

Differential Privacy

Gradients of Deep Neural Networks

Solution to Differentially Private Deep Learning

Abadi et al, 2016

Page 20: Auditing and ML Privacy - Stanford University · Reading Assignments (ML Auditing) Amodei, Dario, Chris Olah, Jacob Steinhardt, Paul Christiano, John Schulman, and Dan Mané. Concrete

Differentially Private SGD

Step 1 Calculate Gradients

Step 3 Adding Noise

Step 4 Parameter Updating

Step 2 Gradient Clipping

Gradient Norm Bounds C

One noise added to each lot (group of data)

Abadi et al, 2016

Page 21: Auditing and ML Privacy - Stanford University · Reading Assignments (ML Auditing) Amodei, Dario, Chris Olah, Jacob Steinhardt, Paul Christiano, John Schulman, and Dan Mané. Concrete

Differentially Private SGD

Abadi et al, 2016

Page 22: Auditing and ML Privacy - Stanford University · Reading Assignments (ML Auditing) Amodei, Dario, Chris Olah, Jacob Steinhardt, Paul Christiano, John Schulman, and Dan Mané. Concrete

Composition Theorem

● If f is (ε1, δ1) - DP (Differential Private) and g is (ε2, δ2) - DP, then

f(D), g(D) is (ε1+ ε2, δ1 + δ2) - DP

Abadi et al, 2016

Page 23: Auditing and ML Privacy - Stanford University · Reading Assignments (ML Auditing) Amodei, Dario, Chris Olah, Jacob Steinhardt, Paul Christiano, John Schulman, and Dan Mané. Concrete

Budget Analysis for Differentially Private SGD

● Bounds the amount of privacy leakage (budget)

● Each lot (group of data) with L samples is (ε, δ) - DP

● Using Composition theorem, our SGD is is (q • ε, q • δ) - DP○ q = L/N - samping ratio per lot

Abadi et al, 2016

Page 24: Auditing and ML Privacy - Stanford University · Reading Assignments (ML Auditing) Amodei, Dario, Chris Olah, Jacob Steinhardt, Paul Christiano, John Schulman, and Dan Mané. Concrete

Moments Accountant

● q = L/N - sampling ratio per lot● T - number of time steps

- Differentially Private

● Provides a tighter bounds for privacy leakage by considering the Gaussian distributed noise

● Under Moments Accountant, there exist c1 and c2 such that Differentially Private SGD is

Abadi et al, 2016

Page 25: Auditing and ML Privacy - Stanford University · Reading Assignments (ML Auditing) Amodei, Dario, Chris Olah, Jacob Steinhardt, Paul Christiano, John Schulman, and Dan Mané. Concrete

ε As A Function of Epoch E● E - number of epochs● q = 0.01● σ = 4● δ = 10−5

Abadi et al, 2016

Page 26: Auditing and ML Privacy - Stanford University · Reading Assignments (ML Auditing) Amodei, Dario, Chris Olah, Jacob Steinhardt, Paul Christiano, John Schulman, and Dan Mané. Concrete

Performance and (ε, δ)

Abadi et al, 2016

Page 27: Auditing and ML Privacy - Stanford University · Reading Assignments (ML Auditing) Amodei, Dario, Chris Olah, Jacob Steinhardt, Paul Christiano, John Schulman, and Dan Mané. Concrete

Performance and Noise Levels

Abadi et al, 2016

σ=8 σ= 4 σ=2

- Differentially Private

Page 28: Auditing and ML Privacy - Stanford University · Reading Assignments (ML Auditing) Amodei, Dario, Chris Olah, Jacob Steinhardt, Paul Christiano, John Schulman, and Dan Mané. Concrete

Outline● ML Auditing

○ Distill-and-Compare

● Privacy in ML○ Differential Privacy with Deep Learning○ Model Inversion Attack and Differential Privacy○ Local Differential Privacy○ Federated Learning

Page 29: Auditing and ML Privacy - Stanford University · Reading Assignments (ML Auditing) Amodei, Dario, Chris Olah, Jacob Steinhardt, Paul Christiano, John Schulman, and Dan Mané. Concrete

Recap: Types of Adversarial Attack

Attack Phase Goal

Evasion Testing Compromise Model Performance

Data Poisoning Training Compromise Model Performance

Exploratory Testing Explore Model CharacteristicsReconstruct User Data

Page 30: Auditing and ML Privacy - Stanford University · Reading Assignments (ML Auditing) Amodei, Dario, Chris Olah, Jacob Steinhardt, Paul Christiano, John Schulman, and Dan Mané. Concrete

Recap● Exploratory Attack

○ Reverse engineer user data from a trained model

Reverse Engineering

Page 31: Auditing and ML Privacy - Stanford University · Reading Assignments (ML Auditing) Amodei, Dario, Chris Olah, Jacob Steinhardt, Paul Christiano, John Schulman, and Dan Mané. Concrete

Model Inversion Attacks

Fredrikson et al, 2015

Original Image Reconstructed Image

Page 32: Auditing and ML Privacy - Stanford University · Reading Assignments (ML Auditing) Amodei, Dario, Chris Olah, Jacob Steinhardt, Paul Christiano, John Schulman, and Dan Mané. Concrete

Model Inversion Attack to Evaluate Differential Privacy

Park et al, 2019

Page 34: Auditing and ML Privacy - Stanford University · Reading Assignments (ML Auditing) Amodei, Dario, Chris Olah, Jacob Steinhardt, Paul Christiano, John Schulman, and Dan Mané. Concrete

Outline● ML Auditing

○ Distill-and-Compare

● Privacy in ML○ Differential Privacy with Deep Learning○ Model Inversion Attack and Differential Privacy○ Local Differential Privacy○ Federated Learning

Page 35: Auditing and ML Privacy - Stanford University · Reading Assignments (ML Auditing) Amodei, Dario, Chris Olah, Jacob Steinhardt, Paul Christiano, John Schulman, and Dan Mané. Concrete

Differential Privacy and Local Differential Privacy

● d, d' are sets of data● d and d' differ in one sample

Differential Privacy

● Centralized setting

Local Differential Privacy

● d and d' are single samples

● Distributed setting

Page 36: Auditing and ML Privacy - Stanford University · Reading Assignments (ML Auditing) Amodei, Dario, Chris Olah, Jacob Steinhardt, Paul Christiano, John Schulman, and Dan Mané. Concrete

Deployment of Local Differential Privacy● RAPPOR by Google

○ Collect user data○ Randomized Aggregatable Privacy-Preserving Ordinal Response

● Private Count Mean Sketch by Apple○ Collect emoji usage data along with other information in iPhone○ Learning with Privacy at Scale

Page 37: Auditing and ML Privacy - Stanford University · Reading Assignments (ML Auditing) Amodei, Dario, Chris Olah, Jacob Steinhardt, Paul Christiano, John Schulman, and Dan Mané. Concrete

Outline● ML Auditing

○ Distill-and-Compare

● Privacy in ML○ Differential Privacy with Deep Learning○ Model Inversion Attack and Differential Privacy○ Local Differential Privacy○ Federated Learning

Page 38: Auditing and ML Privacy - Stanford University · Reading Assignments (ML Auditing) Amodei, Dario, Chris Olah, Jacob Steinhardt, Paul Christiano, John Schulman, and Dan Mané. Concrete

Distributed Optimization

Centralized Setting Distributed Setting

Relies on distributed optimization

Page 39: Auditing and ML Privacy - Stanford University · Reading Assignments (ML Auditing) Amodei, Dario, Chris Olah, Jacob Steinhardt, Paul Christiano, John Schulman, and Dan Mané. Concrete

Federated Optimization ● Non-IID

○ User data is localized to their own usage○ Hard to be a representative of the population

● Unbalanced Similarly○ Some users will make much heavier on particular services than others

● Distributed Computing Capacity○ Expect a large number of devices to be updated at the same time

● Limited communication○ Mobile devices are frequently offline or on slow or expensive connections

McMahan et al, 2017

Page 40: Auditing and ML Privacy - Stanford University · Reading Assignments (ML Auditing) Amodei, Dario, Chris Olah, Jacob Steinhardt, Paul Christiano, John Schulman, and Dan Mané. Concrete

FedSGD

McMahan et al, 2017

GradientModel

Page 41: Auditing and ML Privacy - Stanford University · Reading Assignments (ML Auditing) Amodei, Dario, Chris Olah, Jacob Steinhardt, Paul Christiano, John Schulman, and Dan Mané. Concrete

FedAvg

McMahan et al, 2017

GradientModel

Page 42: Auditing and ML Privacy - Stanford University · Reading Assignments (ML Auditing) Amodei, Dario, Chris Olah, Jacob Steinhardt, Paul Christiano, John Schulman, and Dan Mané. Concrete

Trade-offs Between Local and Global Iterations● Number of rounds of communication necessary to achieve a test-set accuracy

of 97% for the 2NN(MLP) and 99% for the CNN on MNIST

McMahan et al, 2017C - ratio of clients updated to the serverB - batch size of clientsE - number of epochs client makes over its local dataset on each round

Page 43: Auditing and ML Privacy - Stanford University · Reading Assignments (ML Auditing) Amodei, Dario, Chris Olah, Jacob Steinhardt, Paul Christiano, John Schulman, and Dan Mané. Concrete

Comparisons Between FedSGD and FedAvg

McMahan et al, 2017K - number of clientsB - batch size

E - number of epochsu -

Page 44: Auditing and ML Privacy - Stanford University · Reading Assignments (ML Auditing) Amodei, Dario, Chris Olah, Jacob Steinhardt, Paul Christiano, John Schulman, and Dan Mané. Concrete

Effects of Number of Local Epoches

McMahan et al, 2017

Page 45: Auditing and ML Privacy - Stanford University · Reading Assignments (ML Auditing) Amodei, Dario, Chris Olah, Jacob Steinhardt, Paul Christiano, John Schulman, and Dan Mané. Concrete

Effects on η

McMahan et al, 2017

Page 46: Auditing and ML Privacy - Stanford University · Reading Assignments (ML Auditing) Amodei, Dario, Chris Olah, Jacob Steinhardt, Paul Christiano, John Schulman, and Dan Mané. Concrete

Reading Assignments (ML Auditing)● Amodei, Dario, Chris Olah, Jacob Steinhardt, Paul Christiano, John Schulman, and Dan

Mané. Concrete problems in AI safety, arXiv 2016● Kleinberg, Jon, Sendhil Mullainathan, and Manish Raghavan. Inherent trade-offs in the

fair determination of risk scores, arXiv 2016● Malgieri, Gianclaudio. The concept of fairness in the GDPR: a linguistic and contextual

interpretation, FAccT 2020● Goodman, Bryce, and Seth Flaxman. European Union regulations on algorithmic

decision-making and a “right to explanation”, AI magazine 2017● Bellamy, Rachel KE, Kuntal Dey, Michael Hind, Samuel C. Hoffman, Stephanie Houde,

Kalapriya Kannan, Pranay Lohia et al. AI Fairness 360: An extensible toolkit for detecting, understanding, and mitigating unwanted algorithmic bias, arXiv 2018

Page 47: Auditing and ML Privacy - Stanford University · Reading Assignments (ML Auditing) Amodei, Dario, Chris Olah, Jacob Steinhardt, Paul Christiano, John Schulman, and Dan Mané. Concrete

Reading Assignments (Privacy)● Bonawitz, Keith, Vladimir Ivanov, Ben Kreuter, Antonio Marcedone, H. Brendan McMahan, Sarvar

Patel, Daniel Ramage, Aaron Segal, and Karn Seth. Practical secure aggregation for privacy-preserving machine learning, ACM SIGSAC Conference on Computer and Communications Security, 2017

● Konečný, Jakub, H. Brendan McMahan, Felix X. Yu, Peter Richtárik, Ananda Theertha Suresh, and Dave Bacon. Federated learning: Strategies for improving communication efficiency, arXiv 2016

● Bonawitz, Keith, Hubert Eichner, Wolfgang Grieskamp, Dzmitry Huba, Alex Ingerman, Vladimir Ivanov, Chloe Kiddon et al. Towards federated learning at scale: System design, SysML 2019

● Smith, Virginia, Chao-Kai Chiang, Maziar Sanjabi, and Ameet S. Talwalkar. Federated multi-task learning, NeurIPS 2017

● Dwork, Cynthia, Frank McSherry, Kobbi Nissim, and Adam Smith. "Calibrating noise to sensitivity in private data analysis, Theory of cryptography conference 2006