The problem of learning from data while preserving the privacy of individual observations has a long history and spans over multiple disciplines [1–3]. One way to preserve privacy is to corrupt the learning procedure with noise without destroying the information that we want to extract. Differential Privacy (DP) is one of the most powerful tools in this context [3, 4].
Differential Privacy Theory
Oneto L.
2020-01-01
Abstract
The problem of learning from data while preserving the privacy of individual observations has a long history and spans over multiple disciplines [1–3]. One way to preserve privacy is to corrupt the learning procedure with noise without destroying the information that we want to extract. Differential Privacy (DP) is one of the most powerful tools in this context [3, 4].File in questo prodotto:
File | Dimensione | Formato | |
---|---|---|---|
Model+Selection+and+Error+Estimation+in+.pdf
accesso chiuso
Descrizione: Contributo in volume
Tipologia:
Documento in versione editoriale
Dimensione
2.17 MB
Formato
Adobe PDF
|
2.17 MB | Adobe PDF | Visualizza/Apri Richiedi una copia |
I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.