We gratefully acknowledge support from
the Simons Foundation and member institutions.
Full-text links:

Download:

Current browse context:

cs.CL

Change to browse by:

References & Citations

DBLP - CS Bibliography

Bookmark

(what is this?)
CiteULike logo BibSonomy logo Mendeley logo del.icio.us logo Digg logo Reddit logo

Computer Science > Computation and Language

Title: ADELT: Transpilation Between Deep Learning Frameworks

Abstract: We propose the Adversarial DEep Learning Transpiler (ADELT), a novel approach to source-to-source transpilation between deep learning frameworks. ADELT uniquely decouples code skeleton transpilation and API keyword mapping. For code skeleton transpilation, it uses few-shot prompting on large language models (LLMs), while for API keyword mapping, it uses contextual embeddings from a code-specific BERT. These embeddings are trained in a domain-adversarial setup to generate a keyword translation dictionary. ADELT is trained on an unlabeled web-crawled deep learning corpus, without relying on any hand-crafted rules or parallel data. It outperforms state-of-the-art transpilers, improving pass@1 rate by 17.4 pts and 15.0 pts for PyTorch-Keras and PyTorch-MXNet transpilation pairs respectively. We provide open access to our code at this https URL
Comments: 19 pages
Subjects: Computation and Language (cs.CL); Machine Learning (cs.LG)
Cite as: arXiv:2303.03593 [cs.CL]
  (or arXiv:2303.03593v2 [cs.CL] for this version)

Submission history

From: Linyuan Gong [view email]
[v1] Tue, 7 Mar 2023 01:57:10 GMT (234kb,D)
[v2] Mon, 22 Apr 2024 18:18:15 GMT (289kb,D)

Link back to: arXiv, form interface, contact.