Abstract
SingLoRA, a reformulated low-rank adaptation method, enhances parameter-efficient fine-tuning by learning a single low-rank matrix and its transpose, ensuring stable optimization and reducing parameter count.
Low-Rank Adaptation (LoRA) has significantly advanced parameter-efficient fine-tuning of large pretrained models. LoRA augments the pre-trained weights of a model by adding the product of two smaller matrices that together form a low-rank matrix update. Recent research has shown that scale disparities between these two matrices often cause unstable training dynamics, leading to suboptimal performance. In this paper, we propose SingLoRA, which reformulates low-rank adaptation by learning the weights update as a decomposition of a single low-rank matrix multiplied by its transpose. This simple design inherently removes inter-matrix scale conflicts, ensuring stable optimization, and roughly halves the parameter count. We analyze SingLoRA within the infinite-width neural network framework, showing that it guarantees stable feature learning by construction. Extensive experiments on multiple tasks validate these benefits. In common sense reasoning, fine-tuning LLama 7B on MNLI with SingLoRA achieves 91.3% accuracy - surpassing LoRA (89.1%) and LoRA+ (90.2%) - while using only 60% of their parameter budget. In image generation, fine-tuning Stable Diffusion with SingLoRA significantly improves image fidelity on DreamBooth, achieving a DINO similarity score of 0.151, compared to scores of 0.148 and 0.143 for DoRA and LoRA, respectively.
Community
SingLoRA simplifies low-rank adaptation by using a single matrix instead of two, which avoids scale mismatches and reduces parameter count. This simple solution achieves more stable training and exceeds the performance of LoRA and its variants in both language and image tasks.
Made a simple implementation of this paper because there was no github.
arXiv explained breakdown of this paper 👉 https://arxivexplained.com/papers/singlora-low-rank-adaptation-using-a-single-matrix
I would like to know how this method performs compared to LoRI and FFA-LORA (i.e., freezing A and optimizing B).
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- Localized LoRA: A Structured Low-Rank Approximation for Efficient Fine-Tuning (2025)
- Singular Value Decomposition on Kronecker Adaptation for Large Language Model (2025)
- Joint Tensor-Train Parameterization for Efficient and Expressive Low-Rank Adaptation (2025)
- LoRMA: Low-Rank Multiplicative Adaptation for LLMs (2025)
- DiaBlo: Diagonal Blocks Are Sufficient For Finetuning (2025)
- AFLoRA: Adaptive Federated Fine-Tuning of Large Language Models with Resource-Aware Low-Rank Adaption (2025)
- MLorc: Momentum Low-rank Compression for Large Language Model Adaptation (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper