https://arxiv.org/abs/2106.09685
LoRA: Low-Rank Adaptation of Large Language Models
An important paradigm of natural language processing consists of large-scale pre-training on general domain data and adaptation to particular tasks or domains. As we pre-train larger models, full fine-tuning, which retrains all model parameters, becomes le
arxiv.org
참고
'인공지능' 카테고리의 다른 글
GEM Pooling (0) | 2023.12.25 |
---|---|
Segforemer (0) | 2023.12.25 |
Chapter 10 Various artificial intelligence (0) | 2023.05.20 |
Chapter 9 Hyperparameter Optimization (0) | 2023.05.20 |
Chapter 8 Small Nets and EfficientNet (1) | 2023.05.20 |