论文标题

通过加强学习的顺序贝叶斯实验设计

Sequential Bayesian experimental designs via reinforcement learning

论文作者

Asano, Hikaru

论文摘要

贝叶斯实验设计(BED)已被用作基于贝叶斯推断进行有效实验的方法。但是,现有方法主要集中于最大化预期信息增益(EIG);实验和样品效率的成本通常不考虑。为了解决此问题并增强床的实际适用性,我们通过强化学习提供了一种新的方法,通过在本文中应用强化学习,通过增强学习来构建床。在这里,增强学习是机器学习的一个分支,在该分支机构中,代理商通过与环境互动来学习一项政策,以最大程度地提高其奖励。与环境相互作用的特征与顺序实验相似,而增强学习确实是一种在顺序决策中表现出色的方法。 通过提出一个新的面向现实世界的实验环境,我们的方法旨在最大程度地提高特征,同时保持实验成本和样本效率。我们为三个不同示例进行数值实验。可以证实,我们的方法优于各种指数(例如EIG和采样效率)中现有的方法,这表明我们提出的方法和实验环境可以为将床应用于现实世界做出重大贡献。

Bayesian experimental design (BED) has been used as a method for conducting efficient experiments based on Bayesian inference. The existing methods, however, mostly focus on maximizing the expected information gain (EIG); the cost of experiments and sample efficiency are often not taken into account. In order to address this issue and enhance practical applicability of BED, we provide a new approach Sequential Experimental Design via Reinforcement Learning to construct BED in a sequential manner by applying reinforcement learning in this paper. Here, reinforcement learning is a branch of machine learning in which an agent learns a policy to maximize its reward by interacting with the environment. The characteristics of interacting with the environment are similar to the sequential experiment, and reinforcement learning is indeed a method that excels at sequential decision making. By proposing a new real-world-oriented experimental environment, our approach aims to maximize the EIG while keeping the cost of experiments and sample efficiency in mind simultaneously. We conduct numerical experiments for three different examples. It is confirmed that our method outperforms the existing methods in various indices such as the EIG and sampling efficiency, indicating that our proposed method and experimental environment can make a significant contribution to application of BED to the real world.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源