We gratefully acknowledge support from
the Simons Foundation and member institutions.
Full-text links:

Download:

Current browse context:

cs.LG

Change to browse by:

cs

References & Citations

DBLP - CS Bibliography

Bookmark

(what is this?)
CiteULike logo BibSonomy logo Mendeley logo del.icio.us logo Digg logo Reddit logo

Computer Science > Machine Learning

Title: Parameter-Efficient Fine-Tuning for Large Models: A Comprehensive Survey

Abstract: Large models represent a groundbreaking advancement in multiple application fields, enabling remarkable achievements across various tasks. However, their unprecedented scale comes with significant computational costs. These models, often consisting of billions of parameters, require vast amounts of computational resources for execution. Especially, the expansive scale and computational demands pose considerable challenges when customizing them for particular downstream tasks, particularly over the hardware platforms constrained by computational capabilities. Parameter Efficient Fine-Tuning (PEFT) provides a practical solution by efficiently adapt the large models over the various downstream tasks. In particular, PEFT refers to the process of adjusting the parameters of a pre-trained large models to adapt it to a specific task while minimizing the number of additional parameters introduced or computational resources required. This approach is particularly important when dealing with large language models with high parameter counts, as fine-tuning these models from scratch can be computationally expensive and resource-intensive, posing considerable challenges in the supporting system platform design. In this survey, we present comprehensive studies of various PEFT algorithms, examining their performance and computational overhead. Moreover, we provide an overview of applications developed using different PEFT algorithms and discuss common techniques employed to mitigate computation costs for PEFT. In addition to the algorithmic perspective, we overview various real-world system designs to investigate the implementation costs associated with different PEFT algorithms. This survey serves as an indispensable resource for researchers aiming to understand both the PEFT algorithm and its system implementation, offering detailed insights into recent advancements and practical applications.
Comments: 24 pages, 12 figures
Subjects: Machine Learning (cs.LG)
Cite as: arXiv:2403.14608 [cs.LG]
  (or arXiv:2403.14608v5 [cs.LG] for this version)

Submission history

From: Zeyu Han [view email]
[v1] Thu, 21 Mar 2024 17:55:50 GMT (3747kb,D)
[v2] Mon, 1 Apr 2024 15:11:36 GMT (3290kb,D)
[v3] Wed, 17 Apr 2024 16:23:47 GMT (4205kb,D)
[v4] Fri, 26 Apr 2024 13:46:00 GMT (4209kb,D)
[v5] Mon, 29 Apr 2024 16:42:02 GMT (3264kb,D)

Link back to: arXiv, form interface, contact.