Minimal Ranks, Maximum Confidence: Parameter-efficient Uncertainty Quantification for LoRA

Patryk Marszałek, Klaudia Bałazy, Jacek Tabor, Tomasz Kuśmierczyk


Abstract
Low-Rank Adaptation (LoRA) enables parameter-efficient fine-tuning of large language models by decomposing weight updates into low-rank matrices, significantly reducing storage and computational overhead. While effective, standard LoRA lacks mechanisms for uncertainty quantification, leading to overconfident and poorly calibrated models. Bayesian variants of LoRA address this limitation, but at the cost of a significantly increased number of trainable parameters, partially offsetting the original efficiency gains. Additionally, these models are harder to train and may suffer from unstable convergence.In this work, we propose a novel parameter-efficient Bayesian LoRA via subspace inference, demonstrating that effective uncertainty quantification can be achieved in very low-dimensional parameter spaces. The proposed method achieves strong performance with improved calibration and generalization while maintaining computational efficiency. Our empirical findings show that, with the appropriate projection of the weight space: (1) uncertainty can be effectively modeled in a low-dimensional space, and (2) weight covariances exhibit low ranks.
Anthology ID:
2025.findings-emnlp.66
Volume:
Findings of the Association for Computational Linguistics: EMNLP 2025
Month:
November
Year:
2025
Address:
Suzhou, China
Editors:
Christos Christodoulopoulos, Tanmoy Chakraborty, Carolyn Rose, Violet Peng
Venue:
Findings
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
1260–1271
Language:
URL:
https://preview.aclanthology.org/author-page-yu-wang-polytechnic/2025.findings-emnlp.66/
DOI:
10.18653/v1/2025.findings-emnlp.66
Bibkey:
Cite (ACL):
Patryk Marszałek, Klaudia Bałazy, Jacek Tabor, and Tomasz Kuśmierczyk. 2025. Minimal Ranks, Maximum Confidence: Parameter-efficient Uncertainty Quantification for LoRA. In Findings of the Association for Computational Linguistics: EMNLP 2025, pages 1260–1271, Suzhou, China. Association for Computational Linguistics.
Cite (Informal):
Minimal Ranks, Maximum Confidence: Parameter-efficient Uncertainty Quantification for LoRA (Marszałek et al., Findings 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/author-page-yu-wang-polytechnic/2025.findings-emnlp.66.pdf
Checklist:
 2025.findings-emnlp.66.checklist.pdf