TY - JOUR
T1 - Symbiotic Tuning
T2 - A Simple Approach for Enhancing Task Performance of Side-Tuning
AU - Feng, Zhi Quan
AU - Lin, Ying Jia
AU - Chuang, Kun Ta
AU - Kao, Hung-Yu
N1 - Publisher Copyright:
© 2013 IEEE.
PY - 2025
Y1 - 2025
N2 - Reducing computational and memory overhead in fine-tuning large language models remains a significant challenge in natural language processing. While parameter-efficient fine-tuning (PEFT) methods, such as LoRA, have gained attention for reducing trainable parameters while maintaining task performance, they have not achieved substantial memory savings, as memory usage is still dominated by model weights and activations during backpropagation. In contrast, Ladder Side-Tuning (LST) addresses the memory usage problem by freezing the backbone language model (BLM) and training only lightweight side networks. However, this approach often leads to a performance decline compared to PEFT methods. To overcome these limitations, we propose Symbiotic Tuning (SymTune), a novel method that extracts intermediate outputs which includes the hidden states and attention weights from the BLM and integrates symbiotic modules to enhance feature processing capabilities. SymTune strike a much better trade-off between performance and memory efficiency, offering two key advantages: 1) robust performance across a wide range of natural language tasks, and 2) reduced memory consumption through an improved side-tuning architecture. Experimental results demonstrate that SymTune provides a scalable and memory-efficient solution for fine-tuning auto-encoder and auto-regressive language models.
AB - Reducing computational and memory overhead in fine-tuning large language models remains a significant challenge in natural language processing. While parameter-efficient fine-tuning (PEFT) methods, such as LoRA, have gained attention for reducing trainable parameters while maintaining task performance, they have not achieved substantial memory savings, as memory usage is still dominated by model weights and activations during backpropagation. In contrast, Ladder Side-Tuning (LST) addresses the memory usage problem by freezing the backbone language model (BLM) and training only lightweight side networks. However, this approach often leads to a performance decline compared to PEFT methods. To overcome these limitations, we propose Symbiotic Tuning (SymTune), a novel method that extracts intermediate outputs which includes the hidden states and attention weights from the BLM and integrates symbiotic modules to enhance feature processing capabilities. SymTune strike a much better trade-off between performance and memory efficiency, offering two key advantages: 1) robust performance across a wide range of natural language tasks, and 2) reduced memory consumption through an improved side-tuning architecture. Experimental results demonstrate that SymTune provides a scalable and memory-efficient solution for fine-tuning auto-encoder and auto-regressive language models.
UR - https://www.scopus.com/pages/publications/105020887012
UR - https://www.scopus.com/pages/publications/105020887012#tab=citedBy
U2 - 10.1109/ACCESS.2025.3629164
DO - 10.1109/ACCESS.2025.3629164
M3 - Article
AN - SCOPUS:105020887012
SN - 2169-3536
VL - 13
SP - 198471
EP - 198481
JO - IEEE Access
JF - IEEE Access
ER -