Multi-Level Memory for Task Oriented Dialogs
面向任务对话框的多级内存
论文:http://static.tongtianta.site/paper_pdf/279a1bae-eb43-11e9-8f47-00163e08bb86.pdf
Abstract
摘要
Recent end-to-end task oriented dialog systems use memory architectures to incorporate external knowledge in their dialogs. Current work makes simplifying assumptions about the structure of the knowledge base (such as the use of triples to represent knowledge) and combines dialog utterances (context), as well as, knowledge base (KB) results, as part of the same memory. This causes an explosion in the memory size, and makes reasoning over memory, harder. In addition, such a memory design forces hierarchical properties of the data to be fit into a triple structure of memory. This requires the memory reader to learn how to infer relationships across otherwise connected attributes.
最近的端到端任务导向对话系统使用内存体系结构将外部知识纳入其对话中。当前的工作简化了关于知识库结构的假设(例如使用三元组来表示知识),并将对话话语(上下文)以及知识库结果(KB)合并为同一内存的一部分。这会导致内存大小爆炸,并使对内存的推理更加困难。另外,这种存储器设计迫使数据的分层属性适合于存储器的三重结构。这需要内存读取器学习如何推断其他连接属性之间的关系。
In this paper we relax the strong assumptions made by existing architectures and use separate memories for modeling dialog context and KB results. Instead of using triples to store KB results, we introduce a novel multilevel memory architecture consisting of cells for each query and their corresponding results. The multi-level memory first addresses queries, followed by results and finally each key-value pair within a result. We conduct detailed experiments on three publicly available task oriented dialog data sets and we find that our method conclusively outperforms current state-of-the-art models. We report a 15-25% increase in both entity F1 and BLEU scores.
在本文中,我们放宽了现有体系结构的强大假设,并使用单独的内存来建模对话框上下文和KB结果。我们没有使用三元组来存储KB结果,而是引入了一种新颖的多级存储体系结构,该体系结构由每个查询的单元格及其对应的结果组成。多级存储器首先处理查询,然后处理结果,最后处理结果中的每个键值对。我们对三个面向任务的公开对话数据集进行了详细的实验,发现我们的方法最终胜过了当前的最新模型。我们报告,实体F1和BLEU分数均提高了15-25%。
1 Introduction
1引言
Task oriented dialog systems are designed to complete a user specified goal, or service an information request using natural language exchanges. Unlike open domain end-to-end neural dialog models, task oriented systems rely on external knowledge sources, outside of the current conversation context, to return a response (Henderson et al., 2014a; Su et al., 2016; Bordes and Weston, 2017a; Eric and Manning, 2017; El Asri et al., 2017). For instance, in the example shown in Table 1, a dialog agent giving tour package recommendations needs to be able to first query an external knowledge source to determine packages that meet a user’s requirement, and then respond accordingly.
面向任务的对话系统旨在完成用户指定的目标,或使用自然语言交换服务信息请求。与开放域端到端神经对话模型不同,面向任务的系统依赖于当前对话上下文之外的外部知识源来返回响应(Henderson等人,2014a; Su等人,2016; Bordes和Weston ,2017a; Eric和Manning,2017; El Asri等,2017)。例如,在表1所示的示例中,提供旅行套票建议的对话代理程序必须能够首先查询外部知识源,以确定满足用户要求的套票,然后做出相应的响应。
∗Work done during internship at IBM Research AI
∗在IBM Research AI实习期间完成的工作
Table 1: A goal oriented dialog based on the Frames dataset (El Asri et al., 2017) along with an external knowledge source with each row containing a tour package.
表1:基于框架数据集的面向目标的对话(El Asri等人,2017)以及外部知识源,每行包含一个旅游套餐。
In order to enable end-to-end goal oriented dialog tasks, current state of the art methods use neural memory architectures to incorporate external knowledge (Su et al., 2016; Eric and Manning, 2017; Madotto et al., 2018). As can be seen in Table 1, agent responses may also include entity values present only in the dialog context (eg: “Munich” in the Agent response in Turn 2). In order to support such utterances, models also include tokens from the input dialog context in the same memory (Madotto et al., 2018).
为了实现端到端面向目标的对话任务,当前最先进的方法使用神经记忆架构来整合外部知识(Su等人,2016; Eric和Manning,2017; Madotto等人,2018)。如表1所示,代理响应还可以包括仅在对话框上下文中存在的实体值(例如,回合2中代理响应中的“慕尼黑”)。为了支持这种话语,模型还将来自输入对话框上下文的标记包含在同一存储器中(Madotto等,2018)。
所有论文

添加客服微信,加入用户群
蜀ICP备18016327号