TY - GEN
T1 - SBoRA
T2 - 31st International Conference on Neural Information Processing (ICONIP 2024)
AU - Po, Lai-Man
AU - Liu, Yuyang
AU - Wu, Haoxuan
AU - Zhang, Tianqi
AU - Yu, Wing-Yin
AU - Wang, Zhuohan
AU - Jiang, Zeyu
AU - Li, Kun
PY - 2026
Y1 - 2026
N2 - This paper introduces Standard Basis LoRA (SBoRA), a novel parameter-efficient fine-tuning approach for Large Language Models that builds upon the pioneering works of Low-Rank Adaptation (LoRA) and Orthogonal Adaptation. SBoRA reduces the number of trainable parameters by half or doubles the rank with the similar number of trainable parameters as LoRA, while improving learning performance. By utilizing orthogonal standard basis vectors to initialize one of the low-rank matrices (either A or B), SBoRA facilitates regional weight updates and memory-efficient fine-tuning. This results in two variants, SBoRA-FA and SBoRA-FB, where only one of the matrices is updated, leading to a sparse update matrix ΔW with predominantly zero rows or columns. Consequently, most of the fine-tuned model’s weights (W0+ΔW) remain unchanged from the pre-trained weights, akin to the modular organization of the human brain, which efficiently adapts to new tasks. Our empirical results demonstrate the superiority of SBoRA-FA over LoRA in various fine-tuning tasks, including commonsense reasoning and arithmetic reasoning. Furthermore, we evaluate the effectiveness of QSBoRA on quantized LLaMA models of varying scales, highlighting its potential for efficient adaptation to new tasks. Code is available at https://github.com/cityuhkai/SBoRA. © The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2026.
AB - This paper introduces Standard Basis LoRA (SBoRA), a novel parameter-efficient fine-tuning approach for Large Language Models that builds upon the pioneering works of Low-Rank Adaptation (LoRA) and Orthogonal Adaptation. SBoRA reduces the number of trainable parameters by half or doubles the rank with the similar number of trainable parameters as LoRA, while improving learning performance. By utilizing orthogonal standard basis vectors to initialize one of the low-rank matrices (either A or B), SBoRA facilitates regional weight updates and memory-efficient fine-tuning. This results in two variants, SBoRA-FA and SBoRA-FB, where only one of the matrices is updated, leading to a sparse update matrix ΔW with predominantly zero rows or columns. Consequently, most of the fine-tuned model’s weights (W0+ΔW) remain unchanged from the pre-trained weights, akin to the modular organization of the human brain, which efficiently adapts to new tasks. Our empirical results demonstrate the superiority of SBoRA-FA over LoRA in various fine-tuning tasks, including commonsense reasoning and arithmetic reasoning. Furthermore, we evaluate the effectiveness of QSBoRA on quantized LLaMA models of varying scales, highlighting its potential for efficient adaptation to new tasks. Code is available at https://github.com/cityuhkai/SBoRA. © The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2026.
KW - Large Language Models
KW - LoRA
KW - Parameter-Efficient Fine-Tuning
UR - http://www.scopus.com/inward/record.url?scp=105012443725&partnerID=8YFLogxK
UR - https://www.scopus.com/record/pubmetrics.uri?eid=2-s2.0-105012443725&origin=recordpage
U2 - 10.1007/978-981-96-7008-6_28
DO - 10.1007/978-981-96-7008-6_28
M3 - RGC 32 - Refereed conference paper (with host publication)
SN - 9789819670079
T3 - Communications in Computer and Information Science
SP - 387
EP - 401
BT - Neural Information Processing - 31st International Conference, ICONIP 2024, Proceedings, Part XIII
A2 - Mahmud, Mufti
A2 - Doborjeh, Maryam
A2 - Wong, Kevin
A2 - Leung, Andrew Chi Sing
A2 - Doborjeh, Zohreh
A2 - Tanveer, M.
PB - Springer Singapore
Y2 - 2 December 2024 through 6 December 2024
ER -