We gratefully acknowledge support from
the Simons Foundation and member institutions.
Full-text links:

Download:

Current browse context:

cs.LG

Change to browse by:

References & Citations

DBLP - CS Bibliography

Bookmark

(what is this?)
CiteULike logo BibSonomy logo Mendeley logo del.icio.us logo Digg logo Reddit logo

Computer Science > Machine Learning

Title: Stealing the Decoding Algorithms of Language Models

Abstract: A key component of generating text from modern language models (LM) is the selection and tuning of decoding algorithms. These algorithms determine how to generate text from the internal probability distribution generated by the LM. The process of choosing a decoding algorithm and tuning its hyperparameters takes significant time, manual effort, and computation, and it also requires extensive human evaluation. Therefore, the identity and hyperparameters of such decoding algorithms are considered to be extremely valuable to their owners. In this work, we show, for the first time, that an adversary with typical API access to an LM can steal the type and hyperparameters of its decoding algorithms at very low monetary costs. Our attack is effective against popular LMs used in text generation APIs, including GPT-2, GPT-3 and GPT-Neo. We demonstrate the feasibility of stealing such information with only a few dollars, e.g., $\$0.8$, $\$1$, $\$4$, and $\$40$ for the four versions of GPT-3.
Subjects: Machine Learning (cs.LG); Computation and Language (cs.CL); Cryptography and Security (cs.CR)
Journal reference: Proceedings of the 2023 ACM SIGSAC Conference on Computer and Communications Security
DOI: 10.1145/3576915.3616652
Cite as: arXiv:2303.04729 [cs.LG]
  (or arXiv:2303.04729v4 [cs.LG] for this version)

Submission history

From: Ali Naseh [view email]
[v1] Wed, 8 Mar 2023 17:15:58 GMT (3038kb,D)
[v2] Thu, 9 Mar 2023 02:40:44 GMT (4531kb,D)
[v3] Wed, 26 Apr 2023 03:16:43 GMT (2269kb,D)
[v4] Fri, 1 Dec 2023 22:34:34 GMT (2270kb,D)

Link back to: arXiv, form interface, contact.