论文标题

语言模型会窃吗?

Do Language Models Plagiarize?

论文作者

Lee, Jooyoung, Le, Thai, Chen, Jinghui, Lee, Dongwon

论文摘要

过去的文献表明,语言模型(LMS)经常记住培训实例的一部分,并在自然语言生成(NLG)过程中复制它们。但是,尚不清楚LMS在多大程度上“重用”培训语料库。例如,模型可以生成与训练样本相似的释义句子。因此,在这项工作中,我们研究了GPT-2生成的文本中的三种类型的窃(即,逐字,释义和思想)与其训练数据相比,并进一步分析了与域特异性语料库的微调LMS的窃模式,这些模式是在实践中广泛使用的。我们的结果表明,(1)LMS超出记忆的LMS中广泛存在的三种类型的pla窃,(2)LMS的大小和解码方法与它们所表现出的窃程度密切相关,(3)基于它们的相似性和同质性,微型LMS的plagiarism plagiarism模式因其相似性和同质性而异。鉴于大多数LMS的培训数据都是从网络上刮下来的,而无需告知内容所有者,因此将单词,短语甚至从培训集中培训到生成的文本中的核心思想重申,具有道德意义。随着LMS的规模及其培训数据的增加,他们的模式可能会加剧,这引起了人们对使用较大培训语料库进行大型模型的担忧。窃内容还可以包含个人的个人和敏感信息。这些发现总体上对当前LMS在关键任务写作任务中的实用性产生了怀疑,并敦促围绕观察到的现象进行更多讨论。数据和源代码可在https://github.com/brit7777/lm-plagiarism上获得。

Past literature has illustrated that language models (LMs) often memorize parts of training instances and reproduce them in natural language generation (NLG) processes. However, it is unclear to what extent LMs "reuse" a training corpus. For instance, models can generate paraphrased sentences that are contextually similar to training samples. In this work, therefore, we study three types of plagiarism (i.e., verbatim, paraphrase, and idea) among GPT-2 generated texts, in comparison to its training data, and further analyze the plagiarism patterns of fine-tuned LMs with domain-specific corpora which are extensively used in practice. Our results suggest that (1) three types of plagiarism widely exist in LMs beyond memorization, (2) both size and decoding methods of LMs are strongly associated with the degrees of plagiarism they exhibit, and (3) fine-tuned LMs' plagiarism patterns vary based on their corpus similarity and homogeneity. Given that a majority of LMs' training data is scraped from the Web without informing content owners, their reiteration of words, phrases, and even core ideas from training sets into generated texts has ethical implications. Their patterns are likely to exacerbate as both the size of LMs and their training data increase, raising concerns about indiscriminately pursuing larger models with larger training corpora. Plagiarized content can also contain individuals' personal and sensitive information. These findings overall cast doubt on the practicality of current LMs in mission-critical writing tasks and urge more discussions around the observed phenomena. Data and source code are available at https://github.com/Brit7777/LM-plagiarism.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源