Low Kruskal-Rank Adaptation
Yixing Xu ⋅ Guanchen Li ⋅ Chao Li ⋅ Xuanwu Yin ⋅ Dong Li ⋅ Spandan Tiwari ⋅ Ashish Sirasao ⋅ Emad Barsoum
Abstract
Low-rank adaptation (LoRA) is one of the most widely used parameter-efficient fine-tuning (PEFT) methods for adapting pre-trained large language models (LLMs) to downstream tasks. Although LoRA significantly reduces the number of trainable parameters and lowers fine-tuning costs, its performance is often limited by the inherent low-rank assumption. In this paper, we revisit the notion of rank for LoRA update matrices and show that the standard matrix rank fails to capture duplicated directions and redundancy in the update subspace. Motivated by this analysis, we argue that the Kruskal rank offers a more informative criterion for characterizing update diversity. We therefore propose **Low Kruskal Rank Adaptation** (LoKRA), a new PEFT algorithm with provable theoretical guarantees that mitigates the limitations of LoRA. We further introduce LoKRA$^+$, an enhanced variant that provides a tighter theoretical lower bound on the Kruskal rank and yields stronger empirical performance. Experiments on multiple LLMs show that our approach consistently outperforms LoRA and other baselines, establishing state-of-the-art performance across a range of benchmarks.
Successful Page Load