PuSH - Publikationsserver des Helmholtz Zentrums München

Hölzl, F.A.* ; Rueckert, D.* ; Kaissis, G.

Equivariant Differentially Private Deep Learning: Why DP-SGD Needs Sparser Models.

In: (AISec 2023 - Proceedings of the 16th ACM Workshop on Artificial Intelligence and Security, 30 November 2023, Copenhagen, Denmark). 2023. 11-22 (AISec 2023 - Proceedings of the 16th ACM Workshop on Artificial Intelligence and Security)
Verlagsversion DOI
Differentially Private Stochastic Gradient Descent (DP-SGD) limits the amount of private information deep learning models can memorize during training. This is achieved by clipping and adding noise to the model's gradients, and thus networks with more parameters require proportionally stronger perturbation. As a result, large models have difficulties learning useful information, rendering training with DP-SGD exceedingly difficult on more challenging training tasks. Recent research has focused on combating this challenge through training adaptations such as heavy data augmentation and large batch sizes. However, these techniques further increase the computational overhead of DP-SGD and reduce its practical applicability. In this work, we propose using the principle of sparse model design to solve precisely such complex tasks with fewer parameters, higher accuracy, and in less time, thus serving as a promising direction for DP-SGD. We achieve such sparsity by design by introducing equivariant convolutional networks for model training with Differential Privacy. Using equivariant networks, we show that small and efficient architecture design can outperform current state-of-The-Art with substantially lower computational requirements. On CIFAR-10, we achieve an increase of up to 9% in accuracy while reducing the computation time by more than 85%. Our results are a step towards efficient model architectures that make optimal use of their parameters and bridge the privacy-utility gap between private and non-private deep learning for computer vision.
Altmetric
Tags
Anmerkungen
Besondere Publikation
Auf Hompepage verbergern

Zusatzinfos bearbeiten
Eigene Tags bearbeiten
Privat
Eigene Anmerkung bearbeiten
Privat
Auf Publikationslisten für
Homepage nicht anzeigen
Als besondere Publikation
markieren
Publikationstyp Artikel: Konferenzbeitrag
Schlagwörter Designed Sparsity ; Differential Privacy ; Equivariant Convolutions ; Image Classification
Sprache englisch
Veröffentlichungsjahr 2023
HGF-Berichtsjahr 2023
ISSN (print) / ISBN 9798400702600
Konferenztitel AISec 2023 - Proceedings of the 16th ACM Workshop on Artificial Intelligence and Security
Konferzenzdatum 30 November 2023
Konferenzort Copenhagen, Denmark
Quellenangaben Band: , Heft: , Seiten: 11-22 Artikelnummer: , Supplement: ,
Institut(e) Helmholtz Artifical Intelligence Cooperation Unit (HAICU)
Institute for Machine Learning in Biomed Imaging (IML)
POF Topic(s) 30205 - Bioengineering and Digital Health
Forschungsfeld(er) Enabling and Novel Technologies
PSP-Element(e) G-530014-001
G-507100-001
Scopus ID 85179587341
Erfassungsdatum 2024-01-19