论文标题

哪种在开放域多转向对话框,分层或非层次模型中更好?一项实证研究

Which Kind Is Better in Open-domain Multi-turn Dialog,Hierarchical or Non-hierarchical Models? An Empirical Study

论文作者

Lan, Tian, Mao, Xian-Ling, Wei, Wei, Huang, Heyan

论文摘要

目前,开放域的生成对话系统在学术界和行业中引起了极大的关注。尽管单转对话生成成功,但多转向对话的一代仍然是一个巨大的挑战。到目前为止,有两种用于开放域多转向对话框的模型:层次结构和非层次模型。最近,一些作品表明,在其实验环境下,分层模型比非分层模型更好。同时,一些作品也证明了相反的结论。由于缺乏足够的比较,目前尚不清楚哪种模型在开放域的多转向对话框中更好。因此,在本文中,我们将在同一实验设置上系统地测量几乎所有代表性的分层和非分层模型,以检查哪种更好。通过广泛的实验,我们得出以下三个重要结论:(1)除了HRAN模型外,几乎所有分层模型都比开放域多转向对话框的非层次模型差。通过进一步的分析,HRAN的出色表现主要取决于其单词级的注意机制。 (2)如果将单词级别的注意机制整合到这些模型中,其他分层模型的性能也将获得很大的改进。修改后的分层模型甚至显着优于非分层模型。 (3)单词级别的注意机制对于层次模型如此强大的原因是因为它可以更有效地利用上下文信息,尤其是细粒度的信息。此外,我们已经实施了所有模型,并且已经发布了代码。

Currently, open-domain generative dialog systems have attracted considerable attention in academia and industry. Despite the success of single-turn dialog generation, multi-turn dialog generation is still a big challenge. So far, there are two kinds of models for open-domain multi-turn dialog generation: hierarchical and non-hierarchical models. Recently, some works have shown that the hierarchical models are better than non-hierarchical models under their experimental settings; meanwhile, some works also demonstrate the opposite conclusion. Due to the lack of adequate comparisons, it's not clear which kind of models are better in open-domain multi-turn dialog generation. Thus, in this paper, we will measure systematically nearly all representative hierarchical and non-hierarchical models over the same experimental settings to check which kind is better. Through extensive experiments, we have the following three important conclusions: (1) Nearly all hierarchical models are worse than non-hierarchical models in open-domain multi-turn dialog generation, except for the HRAN model. Through further analysis, the excellent performance of HRAN mainly depends on its word-level attention mechanism; (2) The performance of other hierarchical models will also obtain a great improvement if integrating the word-level attention mechanism into these models. The modified hierarchical models even significantly outperform the non-hierarchical models; (3) The reason why the word-level attention mechanism is so powerful for hierarchical models is because it can leverage context information more effectively, especially the fine-grained information. Besides, we have implemented all of the models and already released the codes.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源