Paper 2020/155

Low Latency Privacy-preserving Outsourcing of Deep Neural Network Inference

Yifan Tian, Laurent Njilla, Jiawei Yuan, and Shucheng Yu

Abstract

Efficiently supporting inference tasks of deep neural network (DNN) on the resource-constrained Internet of Things (IoT) devices has been an outstanding challenge for emerging smart systems. To mitigate the burden on IoT devices, one prevalent solution is to outsource DNN inference tasks to the public cloud. However, this type of ``cloud-backed" solutions can cause privacy breach since the outsourced data may contain sensitive information. For privacy protection, the research community has resorted to advanced cryptographic primitives to support DNN inference over encrypted data. Nevertheless, these attempts are limited by the real-time performance due to the heavy IoT computational overhead brought by cryptographic primitives. In this paper, we proposed an edge-computing-assisted framework to boost the efficiency of DNN inference tasks on IoT devices, which also protects the privacy of IoT data to be outsourced. In our framework, the most time-consuming DNN layers are outsourced to edge computing devices. The IoT device only processes compute-efficient layers and fast encryption/decryption. Thorough security analysis and numerical analysis are carried out to show the security and efficiency of the proposed framework. Our analysis results indicate a 99%+ outsourcing rate of DNN operations for IoT devices. Experiments on AlexNet show that our scheme can speed up DNN inference for 40.6X with a 96.2% energy saving for IoT devices.

Metadata
Available format(s)
PDF
Publication info
Preprint. MINOR revision.
Keywords
Deep Neural Network InferencePrivacy-preserving OutsourcingInternet of ThingsEdge Computing
Contact author(s)
jyuan @ umassd edu
History
2020-03-18: revised
2020-02-13: received
See all versions
Short URL
https://ia.cr/2020/155
License
Creative Commons Attribution
CC BY

BibTeX

@misc{cryptoeprint:2020/155,
      author = {Yifan Tian and Laurent Njilla and Jiawei Yuan and Shucheng Yu},
      title = {Low Latency Privacy-preserving Outsourcing of Deep Neural Network Inference},
      howpublished = {Cryptology {ePrint} Archive, Paper 2020/155},
      year = {2020},
      url = {https://eprint.iacr.org/2020/155}
}
Note: In order to protect the privacy of readers, eprint.iacr.org does not use cookies or embedded third party content.