Show simple item record

dc.contributor.authorAgrawal, Rashmi
dc.contributor.authorDe Castro, Leo
dc.contributor.authorJuvekar, Chiraag
dc.contributor.authorChandrakasan, Anantha P
dc.contributor.authorVaikuntanathan, Vinod
dc.contributor.authorJoshi, Ajay
dc.date.accessioned2024-01-03T20:27:09Z
dc.date.available2024-01-03T20:27:09Z
dc.date.issued2023-10-28
dc.identifier.isbn979-8-4007-0329-4
dc.identifier.urihttps://hdl.handle.net/1721.1/153275
dc.description.abstractCloud computing has made it easier for individuals and companies to get access to large compute and memory resources. However, it has also raised privacy concerns about the data that users share with the remote cloud servers. Fully homomorphic encryption (FHE) offers a solution to this problem by enabling computations over encrypted data. Unfortunately, all known constructions of FHE require a noise term for security, and this noise grows during computation. To perform unlimited computations on the encrypted data, we need to perform a periodic noise reduction step known as bootstrapping. This bootstrapping operation is memory-bound as it requires several GBs of data. This leads to orders of magnitude increase in the time required for operating on encrypted data as compared to unencrypted data. In this work, we first present an in-depth analysis of the bootstrapping operation in the CKKS FHE scheme. Similar to other existing works, we observe that CKKS bootstrapping exhibits a low arithmetic intensity (< 1 Op/byte). We then propose memory-aware design (MAD) techniques to accelerate the bootstrapping operation of the CKKS FHE scheme. Our proposed MAD techniques are agnostic of the underlying compute platform and can be equally applied to GPUs, CPUs, FPGAs, and ASICs. Our MAD techniques make use of several caching optimizations that enable maximal data reuse and perform reordering of operations to reduce the amount of data that needs to be transferred to/from the main memory. In addition, our MAD techniques include several algorithmic optimizations that reduce the number of data access pattern switches and the expensive NTT operations. Applying our MAD optimizations for FHE improves bootstrapping arithmetic intensity by 3 ×. For Logistic Regression (LR) training, by leveraging our MAD optimizations, the existing GPU design can get up to 3.5 × improvement in performance for the same on-chip memory size. Similarly, the existing ASIC designs can get up to 27 × and 57 × improvement in performance for LR training and ResNet-20 inference, respectively, while reducing the on-chip memory requirement by 16 ×, which proportionally reduces the cost of the solution.en_US
dc.publisherACM|56th Annual IEEE/ACM International Symposium on Microarchitectureen_US
dc.relation.isversionofhttps://doi.org/10.1145/3613424.3614302en_US
dc.rightsArticle is made available in accordance with the publisher's policy and may be subject to US copyright law. Please refer to the publisher's site for terms of use.en_US
dc.titleMAD: Memory-Aware Design Techniques for Accelerating Fully Homomorphic Encryptionen_US
dc.typeArticleen_US
dc.identifier.citationAgrawal, Rashmi, De Castro, Leo, Juvekar, Chiraag, Chandrakasan, Anantha, Vaikuntanathan, Vinod et al. 2023. "MAD: Memory-Aware Design Techniques for Accelerating Fully Homomorphic Encryption."
dc.contributor.departmentMassachusetts Institute of Technology. Department of Electrical Engineering and Computer Science
dc.identifier.mitlicensePUBLISHER_POLICY
dc.identifier.mitlicensePUBLISHER_POLICY
dc.eprint.versionFinal published versionen_US
dc.type.urihttp://purl.org/eprint/type/ConferencePaperen_US
eprint.statushttp://purl.org/eprint/status/NonPeerRevieweden_US
dc.date.updated2024-01-01T08:47:40Z
dc.language.rfc3066en
dc.rights.holderThe author(s)
dspace.date.submission2024-01-01T08:47:40Z
mit.licensePUBLISHER_POLICY
mit.metadata.statusAuthority Work and Publication Information Neededen_US


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record