$α$-LoRA: Effective Fine-Tuning via Base Model Rescaling
📝 Original Info
- Title: $α$-LoRA: Effective Fine-Tuning via Base Model Rescaling
- ArXiv ID: 2510.21345
- Date: 2025-10-24
- Authors: 정보 없음 (제공된 텍스트에 저자 정보가 포함되어 있지 않습니다.)
📝 Abstract
Fine-tuning has proven to be highly effective in adapting pre-trained models to perform better on new desired tasks with minimal data samples. Among the most widely used approaches are reparameterization methods, which update a target module by augmenting its frozen weight matrix with an additional trainable weight matrix. The most prominent example is Low Rank Adaption (LoRA), which gained significant attention in recent years. In this paper, we introduce a new class of reparameterization methods for transfer learning, designed to enhance the generalization ability of fine-tuned models. We establish the effectiveness of our approach in a high-dimensional binary classification setting using tools from Random Matrix Theory, and further validate our theoretical findings through more realistic experiments, such as fine-tuning LLMs.💡 Deep Analysis
📄 Full Content
Reference
This content is AI-processed based on open access ArXiv data.