论文标题

具有不同隐私的私人微调大语模型

Privately Fine-Tuning Large Language Models with Differential Privacy

论文作者

Behnia, Rouzbeh, Ebrahimi, Mohamamdreza, Pacheco, Jason, Padmanabhan, Balaji

论文摘要

预训练的大语言模型(LLMS)是现代AI不可或缺的一部分,它导致了复杂的AI任务中的突破性表现。具有昂贵基础架构的主要AI公司能够从头开始开发和培训这些大型型号。第三方,研究人员和从业人员越来越多地采用这些预培训的模型,并在其私人数据上进行微调以完成其下游AI任务。但是,已经表明,对手可以从这些LLM中提取/重建精确的培训样本,从而导致揭示个人身份信息。这个问题引起了人们对LLM的隐私的深切关注。差异隐私(DP)提供了一个严格的框架,该框架允许在训练或微调LLMS的过程中添加噪声,从而使提取训练数据变得不可行(即具有密码较小的成功概率)。尽管在大多数现有研究中提供的理论隐私保证通过在渐近环境中的许多培训迭代中从头开始学习模型,但在训练迭代次数少于较小的微调场景中,这种假设并不存在。为了解决差距,我们提出\ ewtune,这是一个基于Edgeworth会计师的微调LLMS的DP框架,并提供有限的样本隐私保证。我们在四种完善的自然语言理解(NLU)任务中的结果表明,尽管\ ewtune〜为LLM微调过程增加了隐私保证,但它直接有助于将诱导的噪声降低到最高5.6 \%,并提高所有NLU任务的最高最高级别的LLMS绩效。我们为广泛采用和公共测试目的开源了我们的实施。

Pre-trained Large Language Models (LLMs) are an integral part of modern AI that have led to breakthrough performances in complex AI tasks. Major AI companies with expensive infrastructures are able to develop and train these large models with billions and millions of parameters from scratch. Third parties, researchers, and practitioners are increasingly adopting these pre-trained models and fine-tuning them on their private data to accomplish their downstream AI tasks. However, it has been shown that an adversary can extract/reconstruct the exact training samples from these LLMs, which can lead to revealing personally identifiable information. The issue has raised deep concerns about the privacy of LLMs. Differential privacy (DP) provides a rigorous framework that allows adding noise in the process of training or fine-tuning LLMs such that extracting the training data becomes infeasible (i.e., with a cryptographically small success probability). While the theoretical privacy guarantees offered in most extant studies assume learning models from scratch through many training iterations in an asymptotic setting, this assumption does not hold in fine-tuning scenarios in which the number of training iterations is significantly smaller. To address the gap, we present \ewtune, a DP framework for fine-tuning LLMs based on Edgeworth accountant with finite-sample privacy guarantees. Our results across four well-established natural language understanding (NLU) tasks show that while \ewtune~adds privacy guarantees to LLM fine-tuning process, it directly contributes to decreasing the induced noise to up to 5.6\% and improves the state-of-the-art LLMs performance by up to 1.1\% across all NLU tasks. We have open-sourced our implementations for wide adoption and public testing purposes.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源