论文标题

KSAT:知识注入的自我关注变压器 - 整合多个领域的上下文

KSAT: Knowledge-infused Self Attention Transformer -- Integrating Multiple Domain-Specific Contexts

论文作者

Roy, Kaushik, Zi, Yuxin, Narayanan, Vignesh, Gaur, Manas, Sheth, Amit

论文摘要

特定于领域的语言理解需要集成多个相关的上下文信息。例如,我们在文本中看到自杀和与抑郁症相关的行为(多个上下文)``我有枪,对自己的生活感到非常难过,如果我明天不醒来,这将不是最糟糕的事情''。自我注意到体系结构中的领域特异性是通过对相关领域特定资源(数据集和外部知识的 - 医学教科书章节的摘录进行微调来处理的,涉及自杀和抑郁症有关)。我们提出了一个经过修改的自我发明体系结构知识注入的自我关注变压器(KSAT),该架构通过使用外部知识来源来实现多个特定领域的环境的整合。 KSAT为每个知识来源引入了专门的自我注意力层中知识引导的偏见,以实现这一目标。此外,KSAT还提供了控制从数据学习和从知识中学习之间权衡的机制。我们的定量和定性评估表明,(1)KSAT体系结构为精确衡量和可视化注入域环境的贡献提供了新颖的人为理解的方法,并且(2)KSAT与其他知识知识的盆地进行竞争性能,并与其他知识的基础线相比,并显着胜过用于domain-specific任务的基准。

Domain-specific language understanding requires integrating multiple pieces of relevant contextual information. For example, we see both suicide and depression-related behavior (multiple contexts) in the text ``I have a gun and feel pretty bad about my life, and it wouldn't be the worst thing if I didn't wake up tomorrow''. Domain specificity in self-attention architectures is handled by fine-tuning on excerpts from relevant domain specific resources (datasets and external knowledge - medical textbook chapters on mental health diagnosis related to suicide and depression). We propose a modified self-attention architecture Knowledge-infused Self Attention Transformer (KSAT) that achieves the integration of multiple domain-specific contexts through the use of external knowledge sources. KSAT introduces knowledge-guided biases in dedicated self-attention layers for each knowledge source to accomplish this. In addition, KSAT provides mechanics for controlling the trade-off between learning from data and learning from knowledge. Our quantitative and qualitative evaluations show that (1) the KSAT architecture provides novel human-understandable ways to precisely measure and visualize the contributions of the infused domain contexts, and (2) KSAT performs competitively with other knowledge-infused baselines and significantly outperforms baselines that use fine-tuning for domain-specific tasks.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源