论文标题
在没有监督的情况下学习稀疏句子编码:探索变异自动编码器中的稀疏性
Learning Sparse Sentence Encoding without Supervision: An Exploration of Sparsity in Variational Autoencoders
论文作者
论文摘要
众所周知,稀疏性是一种有效的归纳偏见,可在具有固定维度的向量中有效地学习数据的有效表示,并且在许多表示学习的领域中都对其进行了探讨。这项工作特别感兴趣的是对VAE框架内的稀疏性进行了调查,该稀疏性在图像域中经过了很多探索,但在NLP中甚至缺乏基本的探索水平。此外,NLP在学习大型文本的稀疏表示方面也落后于句子。我们使用诱导大型文本单位的稀疏潜在表示的VAE来解决上述缺点。首先,我们通过衡量无监督的最先进(SOTA)和其他基于VAE的强大的稀疏基线的成功来朝这个方向移动,并提出了一个分层稀疏VAE模型来解决SOTA的稳定性问题。然后,我们查看稀疏对3个数据集的文本分类的含义,并突出显示下游任务上稀疏潜在表示的性能与其编码与任务相关信息的能力之间的链接。
It has been long known that sparsity is an effective inductive bias for learning efficient representation of data in vectors with fixed dimensionality, and it has been explored in many areas of representation learning. Of particular interest to this work is the investigation of the sparsity within the VAE framework which has been explored a lot in the image domain, but has been lacking even a basic level of exploration in NLP. Additionally, NLP is also lagging behind in terms of learning sparse representations of large units of text e.g., sentences. We use the VAEs that induce sparse latent representations of large units of text to address the aforementioned shortcomings. First, we move in this direction by measuring the success of unsupervised state-of-the-art (SOTA) and other strong VAE-based sparsification baselines for text and propose a hierarchical sparse VAE model to address the stability issue of SOTA. Then, we look at the implications of sparsity on text classification across 3 datasets, and highlight a link between performance of sparse latent representations on downstream tasks and its ability to encode task-related information.