论文标题
迈向个性化的情报
Towards Personalized Intelligence at Scale
论文作者
论文摘要
个性化智能(PI)是为每个用户提供定制的AI体验的问题。在许多应用中,PI是首选甚至需要的。现有的个性化方法涉及对预训练的模型进行微调以创建新的自定义模型。但是,这些方法需要大量的计算来训练,按模型大小和用户数量进行缩放,从而抑制PI被广泛实现。在这项工作中,我们介绍了一种新颖的模型体系结构和培训/推理框架,以使个性化的智能能够大规模。我们通过将个性化头(PH)附加到预训练的语言模型(LM)来实现这一目标。在训练过程中,基本LMS被冷冻,只有pH中的参数已更新,并且每个用户都是唯一的。与传统的微调方法相比,当许多用户缩放时,这会导致总体模型尺寸和培训成本明显小。我们评估了学术界和以行业为中心的数据集的PHS,并表明PHS在F1分数中的表现优于Zeroshot基线,并且比传统的微调方法更可扩展。我们确定有效的pH设计和培训所需的关键因素。
Personalized Intelligence (PI) is the problem of providing customized AI experiences tailored to each individual user. In many applications, PI is preferred or even required. Existing personalization approaches involve fine-tuning pre-trained models to create new customized models. However, these approaches require a significant amount of computation to train, scaling with model size and the number of users, inhibiting PI to be realized widely. In this work, we introduce a novel model architecture and training/inference framework to enable Personalized Intelligence at scale. We achieve this by attaching a Personalization Head (PH) to pre-trained language models (LM). During training, the base LMs are frozen and only the parameters in PH are updated and are unique per user. This results in significantly smaller overall model sizes and training cost than traditional fine-tuning approaches when scaled across many users. We evaluate PHs on academia and industry-focused datasets and show that the PHs outperform zeroshot baseline in F1 score and are significantly more scalable than traditional fine-tuning approaches. We identify key factors required for effective PH design and training.