论文标题
意义的界限:神经机器翻译中的案例研究
The boundaries of meaning: a case study in neural machine translation
论文作者
论文摘要
自然语言处理中深度学习的成功提出了关于语言意义的性质以及自然和人造系统可以处理的方式的有趣问题。一个这样的问题与以来广泛用于语言建模,机器翻译和其他任务中广泛使用的子词分割算法有关。这些算法通常将单词切成语义上不透明的作品,例如'of of'on''','t','t'和ist'in | t | in | t | ist'。然后,该系统代表密集的矢量空间中所产生的段,这有望建模它们之间的语法关系。此表示形式又可以用来将“ | on | t | ist”(英语)映射到'par | od | ont | iste'(法语)。因此,随着学习两种语言的子单词段序列之间的最佳双语映射的任务,翻译的重新概述更高,而不是在词汇层面上进行建模。有时甚至在纯字符序列之间:'p | e | r | i | o | d | o | n | t | t | i | s | t'$ \ rightarrow $'p | a | a | r | r | o | o | o | o | n | t | t | i | s | t | e'。尽管据称具有不透明的性质,但这种子词细分和对齐方式仍在高效的端到端机器翻译系统中工作。此类过程的计算价值是毫无疑问的。但是他们有任何语言或哲学上的合理性吗?我试图通过审查子词分割算法的相关细节,并将它们与重要的哲学和语言辩论联系起来,以使人工智能更加透明和可以解释,从而阐明了这个问题。
The success of deep learning in natural language processing raises intriguing questions about the nature of linguistic meaning and ways in which it can be processed by natural and artificial systems. One such question has to do with subword segmentation algorithms widely employed in language modeling, machine translation, and other tasks since 2016. These algorithms often cut words into semantically opaque pieces, such as 'period', 'on', 't', and 'ist' in 'period|on|t|ist'. The system then represents the resulting segments in a dense vector space, which is expected to model grammatical relations among them. This representation may in turn be used to map 'period|on|t|ist' (English) to 'par|od|ont|iste' (French). Thus, instead of being modeled at the lexical level, translation is reformulated more generally as the task of learning the best bilingual mapping between the sequences of subword segments of two languages; and sometimes even between pure character sequences: 'p|e|r|i|o|d|o|n|t|i|s|t' $\rightarrow$ 'p|a|r|o|d|o|n|t|i|s|t|e'. Such subword segmentations and alignments are at work in highly efficient end-to-end machine translation systems, despite their allegedly opaque nature. The computational value of such processes is unquestionable. But do they have any linguistic or philosophical plausibility? I attempt to cast light on this question by reviewing the relevant details of the subword segmentation algorithms and by relating them to important philosophical and linguistic debates, in the spirit of making artificial intelligence more transparent and explainable.