论文标题
名声:基于功能的对抗性元元素,用于强大的输入表示形式
FAME: Feature-Based Adversarial Meta-Embeddings for Robust Input Representations
论文作者
论文摘要
当不同的嵌入式编码不同的信息时,结合几个嵌入通常会改善下游任务中的性能。已经表明,即使使用来自变压器的嵌入的模型,仍然受益于标准单词嵌入的模型。但是,不同类型和尺寸的嵌入的组合具有挑战性。作为基于注意力的元嵌入的一种替代方案,我们提出了基于特征的对抗元装置(名望),其注意力函数的指导下,该功能以反映单词特异性属性(例如形状和频率)为指导,并表明这有利于处理基于子词的嵌入。此外,FAME使用对抗性训练来优化不同大小的嵌入到同一空间的映射。我们证明,在序列标签和句子分类的跨语言和域上,尤其是在低资源设置中,名声有效地工作。名望为以27种语言,各种NER设置和不同域中的问题分类设置新的最新技术。
Combining several embeddings typically improves performance in downstream tasks as different embeddings encode different information. It has been shown that even models using embeddings from transformers still benefit from the inclusion of standard word embeddings. However, the combination of embeddings of different types and dimensions is challenging. As an alternative to attention-based meta-embeddings, we propose feature-based adversarial meta-embeddings (FAME) with an attention function that is guided by features reflecting word-specific properties, such as shape and frequency, and show that this is beneficial to handle subword-based embeddings. In addition, FAME uses adversarial training to optimize the mappings of differently-sized embeddings to the same space. We demonstrate that FAME works effectively across languages and domains for sequence labeling and sentence classification, in particular in low-resource settings. FAME sets the new state of the art for POS tagging in 27 languages, various NER settings and question classification in different domains.