We gratefully acknowledge support from
the Simons Foundation and member institutions.
Full-text links:

Download:

Current browse context:

cs.LG

Change to browse by:

References & Citations

DBLP - CS Bibliography

Bookmark

(what is this?)
CiteULike logo BibSonomy logo Mendeley logo del.icio.us logo Digg logo Reddit logo

Computer Science > Machine Learning

Title: Deep Classifier Mimicry without Data Access

Abstract: Access to pre-trained models has recently emerged as a standard across numerous machine learning domains. Unfortunately, access to the original data the models were trained on may not equally be granted. This makes it tremendously challenging to fine-tune, compress models, adapt continually, or to do any other type of data-driven update. We posit that original data access may however not be required. Specifically, we propose Contrastive Abductive Knowledge Extraction (CAKE), a model-agnostic knowledge distillation procedure that mimics deep classifiers without access to the original data. To this end, CAKE generates pairs of noisy synthetic samples and diffuses them contrastively toward a model's decision boundary. We empirically corroborate CAKE's effectiveness using several benchmark datasets and various architectural choices, paving the way for broad application.
Comments: 11 pages main, 4 figures, 2 tables, 4 pages appendix
Subjects: Machine Learning (cs.LG); Artificial Intelligence (cs.AI)
Cite as: arXiv:2306.02090 [cs.LG]
  (or arXiv:2306.02090v5 [cs.LG] for this version)

Submission history

From: Steven Braun [view email]
[v1] Sat, 3 Jun 2023 11:45:16 GMT (610kb,D)
[v2] Mon, 11 Mar 2024 14:48:34 GMT (1092kb,D)
[v3] Thu, 21 Mar 2024 09:58:15 GMT (1092kb,D)
[v4] Fri, 12 Apr 2024 11:50:26 GMT (1092kb,D)
[v5] Fri, 26 Apr 2024 06:21:22 GMT (1089kb,D)

Link back to: arXiv, form interface, contact.