Skip to main navigation Skip to search Skip to main content

AROMA: Autonomous Rank-one Matrix Adaptation

Hao Nan Sheng, Zhi-Yong Wang, Hing Cheung So*, Mingrui Yang

*Corresponding author for this work

Research output: Chapters, Conference Papers, Creative and Literary WorksRGC 32 - Refereed conference paper (with host publication)peer-review

3 Downloads (CityUHK Scholars)

Abstract

As large language models continue to grow in size, parameter-efficient fine-tuning (PEFT) has become increasingly crucial. While low-rank adaptation (LoRA) offers a solution through low-rank updates, its static rank allocation may yield suboptimal results. Adaptive low-rank adaptation (AdaLoRA) improves this with dynamic allocation but remains sensitive to initial and target rank configurations. We introduce AROMA, a framework that automatically constructs layer-specific updates by iteratively building up rank-one components with very few trainable parameters that gradually diminish to zero. Unlike existing methods that employ rank reduction mechanisms, AROMA introduces a dual-loop architecture for rank growth. The inner loop extracts information from each rank-one subspace, while the outer loop determines the number of rank-one subspaces, i.e., the optimal rank. We reset optimizer states to maintain subspace independence. AROMA significantly reduces parameters compared to LoRA and AdaLoRA while achieving superior performance on natural language understanding and generation, commonsense reasoning, offering new insights into adaptive PEFT. The code is available at https://github.com/ShuDun23/AROMA.
©2025 Association for Computational Linguistics
Original languageEnglish
Title of host publicationProceedings of the 2025 Conference on Empirical Methods in Natural Language Processing
EditorsChristos Christodoulopoulos, Tanmoy Chakraborty, Carolyn Rose, Violet Peng
PublisherAssociation for Computational Linguistics
Pages3443-3459
Number of pages17
ISBN (Print)979-8-89176-332-6
DOIs
Publication statusPublished - Nov 2025
Event30th Conference on Empirical Methods in Natural Language Processing (EMNLP 2025) - Suzhou, China
Duration: 4 Nov 20259 Nov 2025
https://aclanthology.org/volumes/2025.emnlp-main/

Conference

Conference30th Conference on Empirical Methods in Natural Language Processing (EMNLP 2025)
Abbreviated title30th EMNLP
PlaceChina
CitySuzhou
Period4/11/259/11/25
Internet address

Bibliographical note

Research Unit(s) information for this publication is provided by the author(s) concerned.

Funding

This work was supported by the Research Grant of Shenzhen Research Institute, City University of Hong Kong, Shenzhen, China under Project RIND25501.

Publisher's Copyright Statement

  • This full text is made available under CC-BY 4.0. https://creativecommons.org/licenses/by/4.0/

Fingerprint

Dive into the research topics of 'AROMA: Autonomous Rank-one Matrix Adaptation'. Together they form a unique fingerprint.

Cite this