Differentially Private Optimization on Large Model at Small Cost

Zhiqi Bu, Yu Xiang Wang, Sheng Zha, George Karypis

Research output: Contribution to journalConference articlepeer-review

14 Scopus citations

Abstract

Differentially private (DP) optimization is the standard paradigm to learn large neural networks that are accurate and privacy-preserving. The computational cost for DP deep learning, however, is notoriously heavy due to the per-sample gradient clipping. Existing DP implementations are 2 ∼ 1000× more costly in time and space complexity than the standard (non-private) training. In this work, we develop a novel Book-Keeping (BK) technique that implements existing DP optimizers (thus achieving the same accuracy), with a substantial improvement on the computational cost. Specifically, BK enables DP training on large models and high dimensional data to be roughly as fast and memory-saving as the standard training, whereas previous DP algorithms can be inefficient or incapable of training due to memory error. The computational advantage of BK is supported by the complexity analysis as well as extensive experiments on vision and language tasks. Our implementation achieves state-of-the-art (SOTA) accuracy with very small extra cost: on GPT2 and at almost the same memory cost (< 1% overhead), BK has 1.03× the time complexity of the standard training (0.83× training speed in practice), and 0.61× the time complexity of the most efficient DP implementation (1.36× training speed in practice). We open-source the codebase for the BK algorithm at https://github.com/awslabs/fas t-differential-privacy.

Original languageEnglish (US)
Pages (from-to)3192-3218
Number of pages27
JournalProceedings of Machine Learning Research
Volume202
StatePublished - 2023
Externally publishedYes
Event40th International Conference on Machine Learning, ICML 2023 - Honolulu, United States
Duration: Jul 23 2023Jul 29 2023

Bibliographical note

Publisher Copyright:
© 2023 Proceedings of Machine Learning Research. All rights reserved.

Fingerprint

Dive into the research topics of 'Differentially Private Optimization on Large Model at Small Cost'. Together they form a unique fingerprint.

Cite this