Deep Learning With Differential Privacy Presenter: Xiaojun Xu Deep - - PowerPoint PPT Presentation

deep learning with differential privacy
SMART_READER_LITE
LIVE PREVIEW

Deep Learning With Differential Privacy Presenter: Xiaojun Xu Deep - - PowerPoint PPT Presentation

Deep Learning With Differential Privacy Presenter: Xiaojun Xu Deep Learning Framework Autonomous Driving Gaming Face Recognition Healthcare Deep Learning Framework Dataset Server Model Privacy Issues of Training Data Dataset Server


slide-1
SLIDE 1

Deep Learning With Differential Privacy

Presenter: Xiaojun Xu

slide-2
SLIDE 2

Autonomous Driving Gaming Face Recognition Healthcare

Deep Learning Framework

slide-3
SLIDE 3

Deep Learning Framework

Dataset Server Model

slide-4
SLIDE 4

Privacy Issues of Training Data

Dataset Server Model

slide-5
SLIDE 5

What information will be leaked from the deep learning model?

Dataset Server Model

slide-6
SLIDE 6

Training Privacy Leakage

  • Model Inversion Attack
  • Membership Inference attack
  • Infer whether or not a data case is in the training set.

Model inversion attacks that exploit confidence information and basic countermeasures (CCS’15) Membership inference attacks against machine learning models (Oakland’17)

slide-7
SLIDE 7

Protecting Privacy of Training Data

Dataset Server Model Differential Privacy

slide-8
SLIDE 8

Differential Privacy(DP)

  • Protect the privacy of individuals while allowing

global query.

  • e.g.: how many individuals in the database have

property P?

Individual Property P? … … Alice Yes Victim Yes Bob No … … Individual Property P? … … Alice Yes Bob No … … Database D Database D’ Output of D and D’ should be similar!

slide-9
SLIDE 9

Differential Privacy(DP)

  • Solution: randomize the query answer (e.g. by

adding random noise !.)

Individual Property P? … … Alice Yes Victim Yes Bob No … … Individual Property P? … … Alice Yes Bob No … … 172 +!′ 171 +!

slide-10
SLIDE 10

Differential Privacy

  • Definition: Let A: # → % be a randomized function

database domain # to output domain %. Then A is &, ( -differentially private if for any S ⊆ % and any two databases +, +′ which differs in only one element: Pr / + ∈ 1 ≤ exp & Pr / +6 ∈ 1 + (

slide-11
SLIDE 11

Differentially Private Mechanism

  • Take function !(#), add noise %(#) = ! # + (
  • Noise level is related with:
  • ) and *.
  • The maximal possible difference between !(#) and

!(#′)

slide-12
SLIDE 12

Differentially Private Mechanism

  • Take function !(#), add noise %(#) = ! # + (
  • When adding Gaussian noise, the scale should be:

) = 2 ln(1.25 0 ) Δ2/4

slide-13
SLIDE 13

Pr #$ ∈ & ≤ exp + Pr #, ∈ & + .

Deep Learning with DP

Dataset1 /

1

Dataset2 /

2

slide-14
SLIDE 14

Achieving DP Deep Learning

  • How to achieve DP for deep learning model?
  • Directly adding noise to !.
  • Not releasing model parameters, and during application,

adding noise to the model output.

  • Adding noise in the process training.
slide-15
SLIDE 15

Training Deep Learning Models

  • Model function !

"

  • Training dataset # =

%&, (& , … , % * , ( *

  • Repeat:
  • Sample a batch from #.
  • Learn from the batch by calculating update Δ,.
  • , ≔ , + Δ,
slide-16
SLIDE 16

Deep Learning With DP

slide-17
SLIDE 17

Deep Learning With DP

slide-18
SLIDE 18

What is the bound?

  • At each step the gradient is !, # -DP w.r.t. the

group(batch).

  • What is the DP guarantee after many steps for the

gradient w.r.t. the dataset?

One step, Within the group One step, Within the dataset Many steps, Within the dataset !,#

slide-19
SLIDE 19

Amplification Theorem

  • !: dataset size; ": size of each group
  • # = "/!
  • Amplification Theorem: if gradient is &, ( -DP

within the group, then it’s ) #& , #( -DP within the dataset.

One step, Within the group One step, Within the dataset Many steps, Within the dataset &,( )(#&),#(

slide-20
SLIDE 20

Basic Composition Theorem

  • Applying an !", $" -DP algorithm with an (!&, $&)-

DP algorithm together will give an ( ) !" + !&, $" + $& -DP algorithm.

  • So after ) steps an (!, $)-DP algorithm is ()!, )$)-

DP.

One step, Within the group One step, Within the dataset Many steps, Within the dataset !,$ *(+!),+$ * )+! , )+$

slide-21
SLIDE 21

Strong Composition Theorem

  • Applying the same (", $)-DP algorithm & times will

give an (O " &log

+ ,

, &$)-DP algorithm.

One step, Within the group One step, Within the dataset Many steps, Within the dataset ",$

  • (."),.$
  • ." &log 1/$

, &.$

slide-22
SLIDE 22

Moments Accountant

  • One major contribution of the paper!
  • The noise at each step is Gaussian noise.

One step, Within the group One step, Within the dataset Many steps, Within the dataset !,# $(&!),&# $ &! ( , #

slide-23
SLIDE 23

Comparison

Approach Overall epsilon Overall delta Basic Composition ! "#$ "#% Advanced Composition ! #$ "log 1/% "#% Moments Accountant ! #$ " %

slide-24
SLIDE 24

Experiments

  • MNIST: 70000 gray-level images for hand written

digits with size 28×28.

slide-25
SLIDE 25

Experiments

  • CIFAR10: 60000 colored images of 10 classes with

size 32×32.

slide-26
SLIDE 26

Experiments

  • MNIST: DP-PCA + DP-NeuralNetwork
  • CIFAR-10: Pretrained Conv Layer + DP-NeuralNetwork
slide-27
SLIDE 27

Experiment Results - MNIST

  • Acc without DP: 98.3%
slide-28
SLIDE 28

Experiment Results – CIFAR10

  • Acc without DP: 80%
slide-29
SLIDE 29

Effectiveness

  • What can DP defend?
  • What cannot DP defend?
slide-30
SLIDE 30

Training Privacy Leakage

  • Model Inversion Attack
  • Membership Inference attack
  • Infer whether or not a data case is in the training set.

Model inversion attacks that exploit confidence information and basic countermeasures (CCS’15) Membership inference attacks against machine learning models (Oakland’17)

slide-31
SLIDE 31

What can DP protect?

  • Privacy of individual data in the dataset.
  • Membership Inference Attack
  • Extracting secrets from language models
  • My SSN is “xxx-xx-xxxx”

The Secret Sharer: Measuring Unintended Neural Network Memorization & Extracting Secrets (arXiv preprint)

slide-32
SLIDE 32

What can’t DP protect?

  • Privacy leakage because of global information of

the dataset.

Deep models under the GAN: information leakage from collaborative deep learning (CCS’17)

slide-33
SLIDE 33

Q&A