论文标题
Songnet:刚性格式控制文本生成
SongNet: Rigid Formats Controlled Text Generation
论文作者
论文摘要
神经文本生成在各种任务中取得了巨大进步。大多数任务的一个共同特征是,生成时文本不限于某些刚性格式。但是,我们可能会面对一些特殊的文本范式,例如歌词(假设给出音乐得分),十四行诗,Songci(Songci的古典中国诗歌)等。这些文本的典型特征分为三倍:(1)它们必须完全遵守刚性预定义的预定格式。 (2)他们必须遵守一些押韵计划。 (3)尽管它们仅限于某些格式,但必须保证句子完整性。据我们所知,基于预定义的刚性格式的文本生成尚未得到很好的研究。因此,我们提出了一个名为Songnet的简单而优雅的框架来解决此问题。该框架的骨干是基于变压器的自动回归语言模型。量身定制的符号集以提高建模性能,尤其是在格式,押韵和句子完整性上。我们提高了注意机制,以促进模型以捕获格式的一些未来信息。预培训和微调框架旨在进一步提高发电质量。对两个收集的Corpora进行的广泛实验表明,我们提出的框架在自动指标和人类评估方面都会产生更好的结果。
Neural text generation has made tremendous progress in various tasks. One common characteristic of most of the tasks is that the texts are not restricted to some rigid formats when generating. However, we may confront some special text paradigms such as Lyrics (assume the music score is given), Sonnet, SongCi (classical Chinese poetry of the Song dynasty), etc. The typical characteristics of these texts are in three folds: (1) They must comply fully with the rigid predefined formats. (2) They must obey some rhyming schemes. (3) Although they are restricted to some formats, the sentence integrity must be guaranteed. To the best of our knowledge, text generation based on the predefined rigid formats has not been well investigated. Therefore, we propose a simple and elegant framework named SongNet to tackle this problem. The backbone of the framework is a Transformer-based auto-regressive language model. Sets of symbols are tailor-designed to improve the modeling performance especially on format, rhyme, and sentence integrity. We improve the attention mechanism to impel the model to capture some future information on the format. A pre-training and fine-tuning framework is designed to further improve the generation quality. Extensive experiments conducted on two collected corpora demonstrate that our proposed framework generates significantly better results in terms of both automatic metrics and the human evaluation.