论文标题
感官和学习:无所不在的传感器的自学
Sense and Learn: Self-Supervision for Omnipresent Sensors
论文作者
论文摘要
从无处不在的传感系统(或IoT总体上)从多种传感器数据中学习通用表示,在各种用例中都有许多应用。现有的纯监督的端到端深度学习技术取决于大量策划的数据的可用性,这是众所周知的很难,但需要在利益任务上实现足够的泛化水平。在这项工作中,我们利用自我监督的学习范式来实现从未标记的输入中持续学习的愿景。我们提出了一个名为Sense的广义框架,并学习了从原始感觉数据中的表示或功能学习。它由几个辅助任务组成,这些任务可以从未注释的数据中学习高级和广泛有用的功能,而无需任何人类参与乏味的标签过程。我们证明了我们的方法在来自不同域和各种环境中的几个公开可用数据集上的功效,包括线性可分离性,半监督或少数射击学习以及转移学习。我们的方法可以实现与监督方法竞争的结果,并通过对网络进行微调来缩小差距,同时在大多数情况下学习下游任务。特别是,我们表明可以将自我监督的网络用作初始化,以显着提高低数据表格中的性能,而每个类别的标签实例只有5个标签实例,这对于现实世界中的问题至关重要。同样,即使从目标域可用的标签实例可获得很少的实例,也发现具有自学的学习表示形式在相关数据集之间是高度转移的。我们方法论的自学性质为持续的持续学习打开了令人兴奋的可能性。
Learning general-purpose representations from multisensor data produced by the omnipresent sensing systems (or IoT in general) has numerous applications in diverse use cases. Existing purely supervised end-to-end deep learning techniques depend on the availability of a massive amount of well-curated data, acquiring which is notoriously difficult but required to achieve a sufficient level of generalization on a task of interest. In this work, we leverage the self-supervised learning paradigm towards realizing the vision of continual learning from unlabeled inputs. We present a generalized framework named Sense and Learn for representation or feature learning from raw sensory data. It consists of several auxiliary tasks that can learn high-level and broadly useful features entirely from unannotated data without any human involvement in the tedious labeling process. We demonstrate the efficacy of our approach on several publicly available datasets from different domains and in various settings, including linear separability, semi-supervised or few shot learning, and transfer learning. Our methodology achieves results that are competitive with the supervised approaches and close the gap through fine-tuning a network while learning the downstream tasks in most cases. In particular, we show that the self-supervised network can be utilized as initialization to significantly boost the performance in a low-data regime with as few as 5 labeled instances per class, which is of high practical importance to real-world problems. Likewise, the learned representations with self-supervision are found to be highly transferable between related datasets, even when few labeled instances are available from the target domains. The self-learning nature of our methodology opens up exciting possibilities for on-device continual learning.