[Resource Topic] 2023/1893: BOLT: Privacy-Preserving, Accurate and Efficient Inference for Transformers

Welcome to the resource topic for 2023/1893

Title:
BOLT: Privacy-Preserving, Accurate and Efficient Inference for Transformers

Authors: Qi Pang, Jinhao Zhu, Helen Möllering, Wenting Zheng, Thomas Schneider

Abstract:

The advent of transformers has brought about significant advancements in traditional machine learning tasks. However, their pervasive deployment has raised concerns about the potential leakage of sensitive information during inference. Existing approaches using secure multiparty computation (MPC) face limitations when applied to transformers due to the extensive model size and resource-intensive matrix-matrix multiplications. In this paper, we present BOLT, a privacy-preserving inference framework for transformer models that supports efficient matrix multiplications and nonlinear computations. Combined with our novel machine learning optimizations, BOLT reduces the communication cost by 10.91x. Our evaluation on diverse datasets demonstrates that BOLT maintains comparable accuracy to floating-point models and achieves 4.8-9.5x faster inference across various network settings compared to the state-of-the-art system.

ePrint: https://eprint.iacr.org/2023/1893

See all topics related to this paper.

Feel free to post resources that are related to this paper below.

Example resources include: implementations, explanation materials, talks, slides, links to previous discussions on other websites.

For more information, see the rules for Resource Topics .