Review
Version 1
Preserved in Portico This version is not peer-reviewed
How to Use Transformers for Transfer Learning?
Version 1
: Received: 28 May 2023 / Approved: 31 May 2023 / Online: 31 May 2023 (07:39:24 CEST)
How to cite: Ebrahimzadeh, M.; Asadi, H. How to Use Transformers for Transfer Learning?. Preprints 2023, 2023052185. https://doi.org/10.20944/preprints202305.2185.v1 Ebrahimzadeh, M.; Asadi, H. How to Use Transformers for Transfer Learning?. Preprints 2023, 2023052185. https://doi.org/10.20944/preprints202305.2185.v1
Abstract
Transformers are increasing replacing older generation of deep neural networks due to their success in a wide range of application. The dominant approach of using transformers is to pre-train them on a large training dataset and then fine-tune them on a downstream task. However, as transformers becoming larger, the fine-tuning approach is become an infeasible approach for transfer learning. In this short survey, we list a few recent methods that makes using transformers based on transfer learning more efficient.
Keywords
optics; photonics; light, lasers; journal manuscripts; LaTeX template
Subject
Computer Science and Mathematics, Computer Science
Copyright: This is an open access article distributed under the Creative Commons Attribution License which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
Comments (0)
We encourage comments and feedback from a broad range of readers. See criteria for comments and our Diversity statement.
Leave a public commentSend a private comment to the author(s)
* All users must log in before leaving a comment