论文标题
深入储层计算中的学习效率
Benchmarking Learning Efficiency in Deep Reservoir Computing
论文作者
论文摘要
通常,通过测量测试数据集上的预测能力来评估机器学习模型的性能。这种方法有利于复杂的模型,这些模型可以平稳地拟合复杂的功能并从训练数据点良好地概括。尽管在不同候选模型之间很少报告或比较该学习过程的智力,速度和数据效率的基本组成部分。在本文中,我们介绍了越来越困难的任务以及数据效率指标的基准,以衡量机器学习模型从培训数据中学习的速度。我们将一些已建立的顺序监督模型(例如RNN,LSTMS或Transformers)与基于储层计算的替代模型相对鲜为人知。所提出的任务需要有效解决的广泛计算原始范围,例如内存或计算布尔函数的能力。令人惊讶的是,我们观察到,依靠动态发展的特征地图学习的储层计算系统比具有随机梯度优化训练的完全监督的方法更快,同时实现了可比的精度得分。代码,基准,训练有素的模型以及重现我们的实验的结果,请访问https://github.com/hugcis/benchmark_learning_efficy/。
It is common to evaluate the performance of a machine learning model by measuring its predictive power on a test dataset. This approach favors complicated models that can smoothly fit complex functions and generalize well from training data points. Although essential components of intelligence, speed and data efficiency of this learning process are rarely reported or compared between different candidate models. In this paper, we introduce a benchmark of increasingly difficult tasks together with a data efficiency metric to measure how quickly machine learning models learn from training data. We compare the learning speed of some established sequential supervised models, such as RNNs, LSTMs, or Transformers, with relatively less known alternative models based on reservoir computing. The proposed tasks require a wide range of computational primitives, such as memory or the ability to compute Boolean functions, to be effectively solved. Surprisingly, we observe that reservoir computing systems that rely on dynamically evolving feature maps learn faster than fully supervised methods trained with stochastic gradient optimization while achieving comparable accuracy scores. The code, benchmark, trained models, and results to reproduce our experiments are available at https://github.com/hugcis/benchmark_learning_efficiency/ .