Show simple item record

dc.contributor.authorXiao, Hanshen
dc.contributor.authorSuh, G. Edward
dc.contributor.authorDevadas, Srinivas
dc.date.accessioned2025-01-27T22:32:24Z
dc.date.available2025-01-27T22:32:24Z
dc.date.issued2024-12-02
dc.identifier.isbn979-8-4007-0636-3
dc.identifier.urihttps://hdl.handle.net/1721.1/158081
dc.descriptionCCS ’24, October 14–18, 2024, Salt Lake City, UT, USAen_US
dc.description.abstractWe initiate a formal study on the concept of learnable obfuscation and aim to answer the following question: is there a type of data encoding that maintains the "learnability" of encoded samples, thereby enabling direct model training on transformed data, while ensuring the privacy of both plaintext and the secret encoding function? This long-standing open problem has prompted many efforts to design such an encryption function, for example, NeuraCrypt and TransNet. Nonetheless, all existing constructions are heuristic without formal privacy guarantees, and many successful reconstruction attacks are known on these constructions assuming an adversary with substantial prior knowledge. We present both generic possibility and impossibility results pertaining to learnable obfuscation. On one hand, we demonstrate that any non-trivial, property-preserving transformation which enables effectively learning over encoded samples cannot offer cryptographic computational security in the worst case. On the other hand, from the lens of information-theoretical security, we devise a series of new tools to produce provable and useful privacy guarantees from a set of heuristic obfuscation methods, including matrix masking, data mixing and permutation, through noise perturbation. Under the framework of PAC Privacy, we show how to quantify the leakage from the learnable obfuscation built upon obfuscation and perturbation methods against adversarial inference. Significantly sharpened utility-privacy tradeoffs are achieved compared to state-of-the-art accounting methods when measuring privacy against data reconstruction and membership inference attacks.en_US
dc.publisherACM|Proceedings of the 2024 ACM SIGSAC Conference on Computer and Communications Securityen_US
dc.relation.isversionofhttps://doi.org/10.1145/3658644.3670277en_US
dc.rightsCreative Commons Attributionen_US
dc.rights.urihttps://creativecommons.org/licenses/by/4.0/en_US
dc.sourceAssociation for Computing Machineryen_US
dc.titleFormal Privacy Proof of Data Encoding: The Possibility and Impossibility of Learnable Encryptionen_US
dc.typeArticleen_US
dc.identifier.citationXiao, Hanshen, Suh, G. Edward and Devadas, Srinivas. 2024. "Formal Privacy Proof of Data Encoding: The Possibility and Impossibility of Learnable Encryption."
dc.contributor.departmentMassachusetts Institute of Technology. Department of Electrical Engineering and Computer Scienceen_US
dc.identifier.mitlicensePUBLISHER_CC
dc.eprint.versionFinal published versionen_US
dc.type.urihttp://purl.org/eprint/type/ConferencePaperen_US
eprint.statushttp://purl.org/eprint/status/NonPeerRevieweden_US
dc.date.updated2025-01-01T08:48:23Z
dc.language.rfc3066en
dc.rights.holderThe author(s)
dspace.date.submission2025-01-01T08:48:23Z
mit.licensePUBLISHER_CC
mit.metadata.statusAuthority Work and Publication Information Neededen_US


Files in this item

Thumbnail
Thumbnail

This item appears in the following Collection(s)

Show simple item record