计算机科学
适应(眼睛)
任务(项目管理)
强化学习
利用
集合(抽象数据类型)
秩(图论)
人机交互
分布式计算
多媒体
人工智能
机器学习
计算机安全
程序设计语言
经济
物理
光学
组合数学
管理
数学
作者
Xihe Qiu,Teqi Hao,Shaojie Shi,Xiaoyu Tan,Yu-Jie Xiong
标识
DOI:10.1109/lsp.2024.3377590
摘要
Recently, large language models (LLMs) with conversational-style interaction, such as ChatGPT and Claude, have gained significant importance in the advancement of artificial general intelligence (AGI). However, the extensive resource requirements during pre-training, instruction fine-tuning (IF), and reinforcement learning through human feedback (RLHF) pose challenges, particularly for individuals and studios with limited resources. Moreover, sensitive data that cannot be deployed on remote training platforms or queried through APIs further exacerbates this issue. To address these limitations, researchers have introduced a parameter-efficient framework called low-rank adaptation (LoRA) for IF on LLMs. However, training individual LoRA networks faces capacity constraints and struggles to adapt to large domains with significant distributional shifts across different tasks. In this paper, we propose a novel framework called chain-of-LoRA to enhance the IF performance of LoRA. Our approach involves training a LoRA network to classify the instruction type and then utilizing task-specific LoRA networks to accomplish the respective tasks. By training multiple task-specific LoRA networks, we exploit a trade-off between performance and disk storage, leveraging the easily expandable and cost-effective nature of disk storage compared to precious graphical resources. Our experimental results demonstrate that our proposed framework achieves comparable performance to typical direct IF on LLMs.
科研通智能强力驱动
Strongly Powered by AbleSci AI