News Score: Score the News, Sort the News, Rewrite the Headlines

LoRA vs Full Fine-tuning: An Illusion of Equivalence

View PDF HTML (experimental) Abstract:Fine-tuning is a crucial paradigm for adapting pre-trained large language models to downstream tasks. Recently, methods like Low-Rank Adaptation (LoRA) have been shown to match the performance of fully fine-tuned models on various tasks with an extreme reduction in the number of trainable parameters. Even in settings where both methods learn similarly accurate models, \emph{are their learned solutions really equivalent?} We study how different fine-tuning me...

Read more at arxiv.org

© News Score  score the news, sort the news, rewrite the headlines