While modern machine learning models rely on increasingly large training datasets, data is often limited in privacy-sensitive domains.
Generative models trained with differential privacy (DP) on sensitive data can sidestep this challenge, providing access to synthetic
data instead. However, training DP generative models is highly challenging due to the noise injected into training to enforce DP. We
propose to leverage diffusion models (DMs), an emerging class of deep generative models, and introduce Differentially Private Diffusion
Models (DPDMs), which enforce privacy using differentially private stochastic gradient descent (DP-SGD). We motivate why DP-SGD is well
suited for training DPDMs, and thoroughly investigate the DM parameterization and the sampling algorithm, which turn out to be crucial ingredients
in DPDMs. Furthermore, we propose noise multiplicity, a simple yet powerful modification of the DM training objective tailored to the DP
setting to boost performance. We validate our novel DPDMs on widely-used image generation benchmarks and achieve state-of-the-art (SOTA) performance
by large margins. For example, on MNIST we improve the SOTA FID from 48.4 to 5.01 and downstream classification accuracy from 83.2% to 98.1% for
the privacy setting DP-\((\varepsilon{=}10, \delta{=}10^{-5})\). Moreover, on standard benchmarks, classifiers trained on DPDM-generated synthetic data
perform on par with task-specific DP-SGD-trained classifiers, which has not been demonstrated before for DP generative models.
Modern deep learning usually requires significant amounts of training data. However, sourcing
large datasets in privacy-sensitive domains is often difficult. To circumvent this challenge, generative
models trained on sensitive data can provide access to large synthetic data instead, which can be
used flexibly to train downstream models. Unfortunately, typical overparameterized neural networks
have been shown to provide little to no privacy to the data they have been trained on. For example,
an adversary may be able to recover training images of deep classifiers using gradients of the
networks. Generative models may even overfit directly,
generating data indistinguishable from the
data they have been trained on. In fact, overfitting and privacy-leakage of generative models are more
relevant than ever, considering recent works that train powerful photo-realistic image generators
on large-scale Internet-scraped data. Since the latest variants of these impressive image generation systems
leverage diffusion models, advancing specifically diffusion model-based generative modeling with privacy
guarantees is a pressing topic.
In this work, we propose Differentially Private Diffusion Models (DPDMs), diffusion models (DMs)
trained with rigorous DP guarantees based on differentially private stochastic gradient descent (DP-SGD). Privacy in DP-SGD is enforced by clipping and noising parameter gradients.
We motivate why DMs are uniquely well suited for DP generative modeling, and we study DPDM parameterization, training setting and model sampling in detail,
and optimize it for the DP setup. We propose noise multiplicity to efficiently boost DPDM performance (see Figure above).
Experimentally, we significantly surpass the state-of-the-art in DP synthesis on widely-studied image modeling benchmarks and we demonstrate that classifiers trained
on DPDM-generated data perform on par with task-specific DP-trained discriminative models.
This implies a very high utility of the synthetic data generated by DPDMs, delivering on the promise of DP generative models as an effective data sharing medium.
Key Insights
Experimental Results
We extensively validate our DPDMs on several popular DP benchmark datasets, namely, (conditional) MNIST, (conditional) Fashion-MNIST, and (unconditional) CelebA (downsampled to 32x32 resolution).
We measure sample quality via FID. On MNIST and Fashion-MNIST, we also assess utility of class-labeled generated data by training classifiers on synthesized samples
and compute class prediction accuracy on real data. As is standard practice, we consider logistic regression (Log Reg), MLP, and CNN classifiers; see paper for details.
MNIST & Fashion-MNIST
(Above) MNIST and Fashion-MNIST images generated by existing methods (above black line) and our DPDM (below black line). The DP privacy setting is \((\varepsilon{=}10, \delta{=}10^{-5})\).
(Below) Class-conditional DP image generation performance (MNIST & Fashion-MNIST) measured in FID and downstream classifier utility (\(\delta{=}10^{-5}\) and three \(\varepsilon\)).
DP-MEPF (†) uses additional public data for training (only included for completeness).
CelebA
(Above) CelebA images generated by existing methods (above black line) and our DPDM (below black line). The DP privacy setting is \((\varepsilon{=}10, \delta{=}10^{-6})\).
(Below) DP image generation performance on CelebA measured in FID (\(\delta{=}10^{-6}\) and two \(\varepsilon\)).
G-PATE and DataLens (†) use \(\delta{=}10^{-5}\) (less privacy) and model images at 64x64 resolution.
Paper
Differentially Private Diffusion Models
Tim Dockhorn, Tianshi Cao, Arash Vahdat, Karsten Kreis
@article{dockhorn2022differentially,
title={{Differentially Private Diffusion Models}},
author={Tim Dockhorn and Tianshi Cao and Arash Vahdat and Karsten Kreis},
journal={Transactions on Machine Learning Research},
issn={2835-8856},
year={2023},
url={https://openreview.net/forum?id=ZPpQk7FJXF}
}