论文标题

一个基于图的上下文感知模型,以了解在线对话

A Graph-Based Context-Aware Model to Understand Online Conversations

论文作者

Agarwal, Vibhor, Young, Anthony P., Joglekar, Sagar, Sastry, Nishanth

论文摘要

允许用户参与式参与的在线论坛在公众讨论许多重要问题方面具有变革性。但是,这样的对话有时会升级为仇恨和错误信息的全面交流。自然语言处理(NLP)的现有方法,例如用于分类任务的深度学习模型,仅作为输入单个评论或两次注释,具体取决于任务是否涉及单个注释的属性的推断,或分别在评论对之间的答复。但是,在在线对话中,评论和答复可能基于外部上下文,超出了对模型输入的立即相关信息。因此,意识到对话的环境应改善模型的推理任务的性能。 我们提出了Graphnli,这是一种基于图形的新型深度学习体系结构,它使用图形步道以原则上的方式结合了对话的更广泛背景。具体而言,图形步行从给定的评论开始,并在相同或并行对话线程中示例“附近”评论,这会导致其他嵌入与初始注释的嵌入一起汇总的嵌入。然后,我们将这些丰富的嵌入式用于下游NLP预测任务,这对于在线对话很重要。我们对两个这样的任务进行了评估 - 极性预测和厌恶女性仇恨言语检测 - 发现我们的模型始终优于这两个任务的所有相关基准。具体而言,具有偏见的词根随机步行的Graphnli的宏F1得分分别比在极性预测和仇恨言语检测任务的基于BERT最佳的基线的宏F1分数要好3和6个百分点。

Online forums that allow for participatory engagement between users have been transformative for the public discussion of many important issues. However, such conversations can sometimes escalate into full-blown exchanges of hate and misinformation. Existing approaches in natural language processing (NLP), such as deep learning models for classification tasks, use as inputs only a single comment or a pair of comments depending upon whether the task concerns the inference of properties of the individual comments or the replies between pairs of comments, respectively. But in online conversations, comments and replies may be based on external context beyond the immediately relevant information that is input to the model. Therefore, being aware of the conversations' surrounding contexts should improve the model's performance for the inference task at hand. We propose GraphNLI, a novel graph-based deep learning architecture that uses graph walks to incorporate the wider context of a conversation in a principled manner. Specifically, a graph walk starts from a given comment and samples "nearby" comments in the same or parallel conversation threads, which results in additional embeddings that are aggregated together with the initial comment's embedding. We then use these enriched embeddings for downstream NLP prediction tasks that are important for online conversations. We evaluate GraphNLI on two such tasks - polarity prediction and misogynistic hate speech detection - and found that our model consistently outperforms all relevant baselines for both tasks. Specifically, GraphNLI with a biased root-seeking random walk performs with a macro-F1 score of 3 and 6 percentage points better than the best-performing BERT-based baselines for the polarity prediction and hate speech detection tasks, respectively.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源