Paper 2024/1371

PIGEON: A Framework for Private Inference of Neural Networks

Christopher Harth-Kitzerow, Technical University of Munich, BMW Group
Yongqin Wang, University of Southern California
Rachit Rajat, University of Southern California
Georg Carle
Murali Annavaram, University of Southern California
Abstract

Privacy-Preserving Machine Learning (PPML) is one of the most relevant use cases for Secure Multiparty Computation (MPC). While private training of large neural networks such as VGG-16 or ResNet-50 on state-of-the-art datasets such as ImageNet is still out of reach due to the performance overhead of MPC, GPU-based MPC frameworks are starting to achieve practical runtimes for private inference. However, we show that, in contrast to plaintext machine learning, the usage of GPU acceleration for both linear (e.g., convolutions) and nonlinear neural network layers (e.g., ReLU) is counterproductive in PPML. While GPUs effectively accelerate linear layers compared to CPU-based MPC implementations, the MPC circuits required to evaluate non-linear layers introduce memory overhead and frequent data movement between the GPU and the CPU to handle network communication. This results in slow ReLU performance and high GPU memory requirements in state-of-the-art GPU-based PPML frameworks, hindering them from scaling to multiple images per second inference throughput and batch sizes larger than eight images on ImageNet. To overcome these limitations, we propose PIGEON, an open-source framework for Private Inference of Neural Networks. PIGEON utilizes a novel ABG programming model that switches between arithmetic vectorization and bitslicing on the CPU for non-linear layers, depending on the MPC-specific computation required, while offloading linear layers to the GPU. Compared to the state-of-the-art PPML framework Piranha, PIGEON achieves two orders of magnitude improvements in ReLU throughput, reduces peak GPU memory utilization by one order of magnitude, and scales better with large batch sizes. This translates to one to two orders of magnitude improvements in throughput for large ImageNet batch sizes (e.g., 192) and more than 70% saturation of a 25 Gbit/s network.

Metadata
Available format(s)
PDF
Category
Cryptographic protocols
Publication info
Preprint.
Keywords
Privacy-Preserving Machine LearningSecure InferenceSecure Multiparty ComputationImplementation
Contact author(s)
christopher harth-kitzerow @ tum de
yongqin @ usc edu
rrajt @ usc edu
carle @ net in tum de
annavara @ usc edu
History
2024-12-01: last of 2 revisions
2024-09-01: received
See all versions
Short URL
https://ia.cr/2024/1371
License
Creative Commons Attribution
CC BY

BibTeX

@misc{cryptoeprint:2024/1371,
      author = {Christopher Harth-Kitzerow and Yongqin Wang and Rachit Rajat and Georg Carle and Murali Annavaram},
      title = {{PIGEON}: A Framework for Private Inference of Neural Networks},
      howpublished = {Cryptology {ePrint} Archive, Paper 2024/1371},
      year = {2024},
      url = {https://eprint.iacr.org/2024/1371}
}
Note: In order to protect the privacy of readers, eprint.iacr.org does not use cookies or embedded third party content.