Structure-Aware Low-Rank Adaptation for Parameter-Efficient Fine-Tuning

被引:3
作者
Hu, Yahao [1 ]
Xie, Yifei [1 ]
Wang, Tianfeng [1 ]
Chen, Man [1 ]
Pan, Zhisong [1 ]
机构
[1] Army Engn Univ PLA, Command & Control Engn Coll, Nanjing 210007, Peoples R China
基金
中国国家自然科学基金;
关键词
pre-trained language models; parameter-efficient fine-tuning; low-rank adaptation; intrinsic rank; training efficiency;
D O I
10.3390/math11204317
中图分类号
O1 [数学];
学科分类号
0701 ; 070101 ;
摘要
With the growing scale of pre-trained language models (PLMs), full parameter fine-tuning becomes prohibitively expensive and practically infeasible. Therefore, parameter-efficient adaptation techniques for PLMs have been proposed to learn through incremental updates of pre-trained weights, such as in low-rank adaptation (LoRA). However, LoRA relies on heuristics to select the modules and layers to which it is applied, and assigns them the same rank. As a consequence, any fine-tuning that ignores the structural information between modules and layers is suboptimal. In this work, we propose structure-aware low-rank adaptation (SaLoRA), which adaptively learns the intrinsic rank of each incremental matrix by removing rank-0 components during training. We conduct comprehensive experiments using pre-trained models of different scales in both task-oriented (GLUE) and task-agnostic (Yelp and GYAFC) settings. The experimental results show that SaLoRA effectively captures the structure-aware intrinsic rank. Moreover, our method consistently outperforms LoRA without significantly compromising training efficiency.
引用
收藏
页数:16
相关论文
共 41 条
  • [11] He J., 2022, P INT C LEARN REPR V
  • [12] Heafield K., 2011, P 6 WORKSH STAT MACH, P187
  • [13] Houlsby N, 2019, PR MACH LEARN RES, V97
  • [14] Hu E.J., 2022, P INT C LEARN REPR V
  • [15] Kossen J, 2024, Arxiv, DOI [arXiv:2307.12375, DOI 10.48550/ARXIV.2307.12375]
  • [16] Lee Jaejun, 2019, arXiv
  • [17] Lester B, 2021, 2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), P3045
  • [18] Li C., 2018, P INT C LEARN REPR V
  • [19] Li J, 2018, P 2018 C N AM CHAPTE, V1, P1865, DOI [10.18653/v1/n18-1169, DOI 10.18653/V1/N18-1169]
  • [20] Li XLS, 2021, 59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (ACL-IJCNLP 2021), VOL 1, P4582