一种面向跨域场景的隐私保护大语言模型微调方法
首发时间:2025-02-26
摘要:大语言模型是基于深度学习的自然语言处理模型,通过领域数据微调可提升特定任务表现。然而,中小企业往往不具备预训练大模型的能力,需要和其他大模型持有者合作进行跨域微调。同时,微调数据中可能包含隐私信息,直接使用会导致隐私信息泄露。目前,相关研究大多是通过由微调数据持有方向预训练大模型持有方发送满足差分隐私的微调数据以实现隐私保护的跨域微调,但这样微调得到的模型将不受到微调数据提供方的控制。此外,差分隐私需要对嵌入加入随机噪音,这会降低模型的效用。因此,本文提出了一种面向跨域场景的隐私保护大语言模型微调方法,PTCDS。在该方法中,针对微调得到的模型不受微调数据提供方控制的问题,设计了基于拆分学习的隐私保护跨域微调方法;针对差分隐私导致的模型效用降低问题,设计了基于嵌入可用性的参数更新降噪方法。实验结果表明,与现有相关工作相比,PTCDS在能够有效保护微调数据的隐私的同时,具有较高的精度。
关键词: 人工智能 隐私保护 差分隐私 大语言模型 提示微调
For information in English, please click here
A Privacy-preserving LLM Fine-tuning Method for Cross-domain Scenarios
Abstract:Large language models are natural language processing models based on deep learning, and their performance on specific tasks can be enhanced through fine-tuning with domain-specific data. However, small and medium-sized enterprises often lack the capability to pre-train large models and need to collaborate with other large model owners to conduct cross-domain fine-tuning. Meanwhile, the fine-tuning data may contain privacy information, and its direct use could lead to privacy leakage. Currently, most relevant studies achieve cross-domain fine-tuning with privacy protection by having the fine-tuning data holder send fine-tuning data satisfying differential privacy to the pre-trained large model holder, but the fine-tuned model obtained in this way will not be controlled by the fine-tuning data provider. Furthermore, differential privacy requires adding random noise to the embeddings, which can reduce the model\'s utility. Therefore, this paper proposes a privacy-preserving large language model fine-tuning method for cross-domain scenarios, PTCDS. In this method, to address the issue that the fine-tuned model is not controlled by the fine-tuning data provider, a privacy-preserving cross-domain fine-tuning framework based on split learning is designed; to address the issue of reduced model utility caused by differential privacy, a parameter update denoising method based on embedding availability is designed. Experimental results show that, compared with existing relevant work, PTCDS can effectively protect the privacy of fine-tuning data while achieving high accuracy.
Keywords: Artificial Intelligence Privacy Protection Differential Privacy Large Language Model Prompt Tuning
基金:
引用
No.****
同行评议
勘误表
一种面向跨域场景的隐私保护大语言模型微调方法
评论
全部评论