论文标题
使大型语言模型交互式:一项关于支持具有隐式限制的复杂信息寻求信息任务的先驱研究
Making Large Language Models Interactive: A Pioneer Study on Supporting Complex Information-Seeking Tasks with Implicit Constraints
论文作者
论文摘要
当前具有自然语言界面的交互式系统无法理解复杂的信息寻求请求,该请求一次表达了几个隐式约束,并且没有关于用户偏好的先前信息,例如,在旧金山周围找到远足踪迹,可与Toddlers访问,可以在夏季使用夏季的风景美丽的风景”,以便他们可以启动探索用户的列表,以供他们求解使用。在这种情况下,与对话和探索性搜索模型不同,可以以复杂且较长的查询形式以一个镜头发出用户请求,在这种情况下,通常需要简短的话语或查询,通常会逐步向系统呈现。我们已经设计并部署了一个平台,以收集接近这种复杂交互系统的数据。此外,尽管当前有生成语言模型的进步,这些模型在提供准确的事实知识方面却遭受了幻觉。所有语言模型大部分都在过去的Web网络数据中很大程度上培训,这通常对直接用户的需求没有用。在本文中,我们提出了一个利用大型语言模型(LLM)进行复杂要求理解的IA,并使用加固学习使其具有互动性,从而使用户的请求通过使其完整,从而使其完善,从而更好地检索并减少LLMS幻觉问题,以满足当前用户需求。为了证明所提出的建模范式的性能,我们采用了各种反应前指标,以捕获引导与系统相互作用的程度,从而获得更好的检索结果。通过广泛的实验,我们证明了我们的方法显着优于几个强大的基准。
Current interactive systems with natural language interfaces lack the ability to understand a complex information-seeking request which expresses several implicit constraints at once, and there is no prior information about user preferences e.g.,"find hiking trails around San Francisco which are accessible with toddlers and have beautiful scenery in summer", where output is a list of possible suggestions for users to start their exploration. In such scenarios, user requests can be issued in one shot in the form of a complex and long query, unlike conversational and exploratory search models, where require short utterances or queries are often presented to the system step by step. We have designed and deployed a platform to collect the data from approaching such complex interactive systems. Moreover, despite with the current advancement of generative language models these models suffer from hallucination in providing accurate factual knowledge. All language models are mostly trained in large part on web-scraped data from the past, which usually is not useful for immediate users' needs. In this article, we propose an IA that leverages Large Language Models (LLM) for complex request understanding and makes it interactive using Reinforcement learning that allows intricately refine user requests by making them complete, leading to better retrieval and reduce LLMs hallucination problems for current user needs. To demonstrate the performance of the proposed modeling paradigm, we have adopted various pre-retrieval metrics that capture the extent to which guided interactions with our system yield better retrieval results. Through extensive experimentation, we demonstrated that our method significantly outperforms several robust baselines.