@inproceedings{cherniuk-etal-2025-run,
title = "Run {L}o{RA} Run: Faster and Lighter {L}o{RA} Implementations",
author = "Cherniuk, Daria and
Mikhalev, Aleksandr and
Oseledets, Ivan",
editor = "Rehm, Georg and
Li, Yunyao",
booktitle = "Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 6: Industry Track)",
month = jul,
year = "2025",
address = "Vienna, Austria",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/display_plenaries/2025.acl-industry.15/",
pages = "200--207",
ISBN = "979-8-89176-288-6",
abstract = "LoRA is a technique that reduces the number of trainable parameters in a neural network by introducing low-rank adapters to linear layers. This technique is used for fine-tuning and even training large transformer models from scratch. This paper presents the RunLoRA framework for efficient implementations of LoRA, which significantly improves the speed of neural network training and fine-tuning with low-rank adapters. The proposed implementation optimizes the computation of LoRA operations based on the shape of the corresponding linear layer weights, the input dimensions, and the LoRA rank by selecting the best forward and backward computation graphs based on FLOPs and time estimations. This results in faster training without sacrificing accuracy. The experimental results show a speedup ranging from 10{\%} to 28{\%} on various transformer models."
}
Markdown (Informal)
[Run LoRA Run: Faster and Lighter LoRA Implementations](https://preview.aclanthology.org/display_plenaries/2025.acl-industry.15/) (Cherniuk et al., ACL 2025)
ACL
- Daria Cherniuk, Aleksandr Mikhalev, and Ivan Oseledets. 2025. Run LoRA Run: Faster and Lighter LoRA Implementations. In Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 6: Industry Track), pages 200–207, Vienna, Austria. Association for Computational Linguistics.