论文标题
建模建议对准问题
Modelling the Recommender Alignment Problem
论文作者
论文摘要
推荐系统(RS)在线调节人类体验。大多数RS ACT是优化与最佳用户不完全一致但易于衡量的指标,例如广告单击和用户参与度。这导致了许多难以估量的副作用:政治两极分化,成瘾,假新闻。 RS设计面临着一个建议的对齐问题:将建议与用户,系统设计师和整个社会的目标保持一致。但是,我们如何测试和比较潜在的解决方案以对齐Rs?他们的规模使他们在部署中进行测试的昂贵和风险。我们合成了一个简单的抽象建模框架来指导未来的工作。 为了说明它,我们构建了一个玩具实验,我们在其中问:“我们如何评估使用用户保留作为奖励功能的后果?”为了回答这个问题,我们学习通过在玩具环境上控制图形动力学来优化奖励功能的建议策略。根据培训推荐人对环境的影响,我们得出结论,最大化者通常会导致比对齐的推荐人更糟糕的结果,但并非总是如此。学习后,我们将RS之间的竞争作为RS对齐的潜在解决方案。我们发现,这通常使我们的玩具社会变得更好,而不是在没有建议或最大化器的情况下。 在这项工作中,我们旨在建立广泛的范围,从表面上触摸许多不同的角度,以阐明如何对推荐系统进行奖励功能的端到端研究。建议对齐是一个紧迫而重要的问题。尝试的解决方案肯定会产生深远的影响。在这里,我们迈出了开发方法来评估和比较解决方案对社会的影响的第一步。
Recommender systems (RS) mediate human experience online. Most RS act to optimize metrics that are imperfectly aligned with the best-interest of users but are easy to measure, like ad-clicks and user engagement. This has resulted in a host of hard-to-measure side-effects: political polarization, addiction, fake news. RS design faces a recommender alignment problem: that of aligning recommendations with the goals of users, system designers, and society as a whole. But how do we test and compare potential solutions to align RS? Their massive scale makes them costly and risky to test in deployment. We synthesized a simple abstract modelling framework to guide future work. To illustrate it, we construct a toy experiment where we ask: "How can we evaluate the consequences of using user retention as a reward function?" To answer the question, we learn recommender policies that optimize reward functions by controlling graph dynamics on a toy environment. Based on the effects that trained recommenders have on their environment, we conclude that engagement maximizers generally lead to worse outcomes than aligned recommenders but not always. After learning, we examine competition between RS as a potential solution to RS alignment. We find that it generally makes our toy-society better-off than it would be under the absence of recommendation or engagement maximizers. In this work, we aimed for a broad scope, touching superficially on many different points to shed light on how an end-to-end study of reward functions for recommender systems might be done. Recommender alignment is a pressing and important problem. Attempted solutions are sure to have far-reaching impacts. Here, we take a first step in developing methods to evaluating and comparing solutions with respect to their impacts on society.