CRYPTOGRU: Low Latency Privacy-Preserving Text Analysis With GRU

Bo Feng, Qian Lou, Lei Jiang, Geoffrey Fox


Abstract
Homomorphic encryption (HE) and garbled circuit (GC) provide the protection for users’ privacy. However, simply mixing the HE and GC in RNN models suffer from long inference latency due to slow activation functions. In this paper, we present a novel hybrid structure of HE and GC gated recurrent unit (GRU) network, , for low-latency secure inferences. replaces computationally expensive GC-based tanh with fast GC-based ReLU, and then quantizes sigmoid and ReLU to smaller bit-length to accelerate activations in a GRU. We evaluate with multiple GRU models trained on 4 public datasets. Experimental results show achieves top-notch accuracy and improves the secure inference latency by up to 138× over one of the state-of-the-art secure networks on the Penn Treebank dataset.
Anthology ID:
2021.emnlp-main.156
Volume:
Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing
Month:
November
Year:
2021
Address:
Online and Punta Cana, Dominican Republic
Venue:
EMNLP
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
2052–2057
Language:
URL:
https://aclanthology.org/2021.emnlp-main.156
DOI:
10.18653/v1/2021.emnlp-main.156
Bibkey:
Cite (ACL):
Bo Feng, Qian Lou, Lei Jiang, and Geoffrey Fox. 2021. CRYPTOGRU: Low Latency Privacy-Preserving Text Analysis With GRU. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pages 2052–2057, Online and Punta Cana, Dominican Republic. Association for Computational Linguistics.
Cite (Informal):
CRYPTOGRU: Low Latency Privacy-Preserving Text Analysis With GRU (Feng et al., EMNLP 2021)
Copy Citation:
PDF:
https://preview.aclanthology.org/ingestion-script-update/2021.emnlp-main.156.pdf
Video:
 https://preview.aclanthology.org/ingestion-script-update/2021.emnlp-main.156.mp4
Data
IMDb Movie Reviews