tCURLoRA: tensor CUR decomposition based low-rank parameter adaptation and its application in medical image segmentation
tCURLoRA: tensor CUR decomposition based low-rank parameter adaptation and its application in medical image segmentation
Transfer learning, by leveraging knowledge from pre-trained models, has significantly enhanced the performance of target tasks. However, as deep neural networks scale up, full fine-tuning introduces substantial computational and storage challenges in resource-constrained environments, limiting its widespread adoption. To address this, parameter-efficient fine-tuning (PEFT) methods have been developed to reduce computational complexity and storage requirements by minimizing the number of updated parameters. While matrix decomposition-based PEFT methods, such as LoRA, show promise, they struggle to fully capture the high-dimensional structural characteristics of model weights. In contrast, high-dimensional tensors offer a more natural representation of neural network weights, allowing for a more comprehensive capture of higher-order features and multi-dimensional interactions. In this paper, we propose tCURLoRA, a novel fine-tuning method based on tensor CUR decomposition. By concatenating pre-trained weight matrices into a three-dimensional tensor and applying tensor CUR decomposition, we update only the lower-order tensor components during fine-tuning, effectively reducing computational and storage overhead. Experimental results demonstrate that tCURLoRA outperforms existing PEFT methods in medical image segmentation tasks.
eess.IV, cs.CV
He, Guanghua
61410eb5-0163-4a67-b691-2a350d0b64a2
Cheng, Wangang
e1d5dea3-608f-4cd2-ad5f-82e15021ddf6
Zhu, Hancan
494e4ce9-6ccf-4ab4-9beb-936dd2686f93
Cai, Xiaohao
de483445-45e9-4b21-a4e8-b0427fc72cee
Yu, Gaohang
304aa1da-05ff-48d0-a19d-1e3a74de273b
4 January 2025
He, Guanghua
61410eb5-0163-4a67-b691-2a350d0b64a2
Cheng, Wangang
e1d5dea3-608f-4cd2-ad5f-82e15021ddf6
Zhu, Hancan
494e4ce9-6ccf-4ab4-9beb-936dd2686f93
Cai, Xiaohao
de483445-45e9-4b21-a4e8-b0427fc72cee
Yu, Gaohang
304aa1da-05ff-48d0-a19d-1e3a74de273b
[Unknown type: UNSPECIFIED]
Abstract
Transfer learning, by leveraging knowledge from pre-trained models, has significantly enhanced the performance of target tasks. However, as deep neural networks scale up, full fine-tuning introduces substantial computational and storage challenges in resource-constrained environments, limiting its widespread adoption. To address this, parameter-efficient fine-tuning (PEFT) methods have been developed to reduce computational complexity and storage requirements by minimizing the number of updated parameters. While matrix decomposition-based PEFT methods, such as LoRA, show promise, they struggle to fully capture the high-dimensional structural characteristics of model weights. In contrast, high-dimensional tensors offer a more natural representation of neural network weights, allowing for a more comprehensive capture of higher-order features and multi-dimensional interactions. In this paper, we propose tCURLoRA, a novel fine-tuning method based on tensor CUR decomposition. By concatenating pre-trained weight matrices into a three-dimensional tensor and applying tensor CUR decomposition, we update only the lower-order tensor components during fine-tuning, effectively reducing computational and storage overhead. Experimental results demonstrate that tCURLoRA outperforms existing PEFT methods in medical image segmentation tasks.
Text
2501.02227v2
- Author's Original
More information
Published date: 4 January 2025
Keywords:
eess.IV, cs.CV
Identifiers
Local EPrints ID: 497986
URI: http://eprints.soton.ac.uk/id/eprint/497986
PURE UUID: e850de0c-666a-4b7c-a512-5b27cc8725cf
Catalogue record
Date deposited: 05 Feb 2025 18:06
Last modified: 06 Feb 2025 03:01
Export record
Altmetrics
Contributors
Author:
Guanghua He
Author:
Wangang Cheng
Author:
Hancan Zhu
Author:
Xiaohao Cai
Author:
Gaohang Yu
Download statistics
Downloads from ePrints over the past year. Other digital versions may also be available to download e.g. from the publisher's website.
View more statistics