论文标题
通过时间频率一致性为时间序列进行自我监督的对比预训练
Self-Supervised Contrastive Pre-Training For Time Series via Time-Frequency Consistency
论文作者
论文摘要
预先训练时间序列构成了独特的挑战,这是由于预训练和目标域之间的潜在不匹配,例如时间动态的变化,快速变化的趋势以及远程和短期效应,这可能会导致下游性能不佳。尽管域适应方法可以减轻这些偏移,但大多数方法都需要直接从目标域中进行示例,从而使其次优于预训练。为了应对这一挑战,方法需要适应具有不同时间动力学的目标域,并且能够在预训练期间看到任何目标示例。相对于其他方式,在时间序列中,我们期望同一示例的基于时间和频率的表示形式位于时间频空间中。为此,我们认为时间频一致性(TF-C)(嵌入了基于频率基于频率的社区的示例的基于时间的社区)是可取的。在TF-C的启发下,我们定义了一个可分解的预训练模型,其中自我监管的信号由时间和频率分量之间的距离提供,每个信号通过对比度估计单独训练。我们在八个数据集上评估了新方法,包括电诊断测试,人类活动识别,机械故障检测和身体状态监测。针对八种最先进方法的实验表明,TF-C在一对一的环境中平均超过15.4%(F1得分)(例如,在EMG数据上对EEG预测的模型进行微调),在挑战一对偏远的设置方面(例如,对EEG的模型)进行了8.4%(精度)(精确),以供您使用eeg的模型(例如,均一范围的模型)现实世界应用中出现的场景广度。代码和数据集:https://github.com/mims-harvard/tfc-pretraining。
Pre-training on time series poses a unique challenge due to the potential mismatch between pre-training and target domains, such as shifts in temporal dynamics, fast-evolving trends, and long-range and short-cyclic effects, which can lead to poor downstream performance. While domain adaptation methods can mitigate these shifts, most methods need examples directly from the target domain, making them suboptimal for pre-training. To address this challenge, methods need to accommodate target domains with different temporal dynamics and be capable of doing so without seeing any target examples during pre-training. Relative to other modalities, in time series, we expect that time-based and frequency-based representations of the same example are located close together in the time-frequency space. To this end, we posit that time-frequency consistency (TF-C) -- embedding a time-based neighborhood of an example close to its frequency-based neighborhood -- is desirable for pre-training. Motivated by TF-C, we define a decomposable pre-training model, where the self-supervised signal is provided by the distance between time and frequency components, each individually trained by contrastive estimation. We evaluate the new method on eight datasets, including electrodiagnostic testing, human activity recognition, mechanical fault detection, and physical status monitoring. Experiments against eight state-of-the-art methods show that TF-C outperforms baselines by 15.4% (F1 score) on average in one-to-one settings (e.g., fine-tuning an EEG-pretrained model on EMG data) and by 8.4% (precision) in challenging one-to-many settings (e.g., fine-tuning an EEG-pretrained model for either hand-gesture recognition or mechanical fault prediction), reflecting the breadth of scenarios that arise in real-world applications. Code and datasets: https://github.com/mims-harvard/TFC-pretraining.