Back to Search Start Over

LLMs Can Understand Encrypted Prompt: Towards Privacy-Computing Friendly Transformers

Authors :
Liu, Xuanqi
Liu, Zhuotao
Publication Year :
2023

Abstract

The community explored to build private inference frameworks for transformer-based large language models (LLMs) in a server-client setting, where the server holds the model parameters and the client inputs its private data (or prompt) for inference. However, these frameworks impose significant overhead when the private inputs are forward propagated through the original LLMs. In this paper, we show that substituting the computation- and communication-heavy operators in the transformer architecture with privacy-computing friendly approximations can greatly reduce the private inference costs while incurring very minor impact on model performance. Compared to state-of-the-art Iron (NeurIPS 2022), our privacy-computing friendly model inference pipeline achieves a $5\times$ acceleration in computation and an 80% reduction in communication overhead, while retaining nearly identical accuracy.

Details

Database :
arXiv
Publication Type :
Report
Accession number :
edsarx.2305.18396
Document Type :
Working Paper