论文标题
用于增强多语言知识和文本建模的适配器
Adapters for Enhanced Modeling of Multilingual Knowledge and Text
论文作者
论文摘要
大型语言模型似乎从他们接受过的大型文本语料库中学习了事实。这些事实是在其许多参数中隐含地编码的,因此很难验证或操纵所学知识。语言模型最近已扩展到多语言语言模型(MLLM),从而使知识可以从数百种语言中学习。同时,知识图包含明确的三重格式的事实,需要仔细且昂贵的策划,并且只有几种高资源语言才能提供,从而限制了他们的研究和应用。为了解决这些问题,我们建议通过来自多语言知识图(MLKG)的知识来增强MLLM,以解决许多语言(包括低资源的语言)的语言和知识图形任务。具体而言,我们引入了一个轻巧的适配器集,以增强MLLM,并通过跨语性实体对齐和MLKGS的事实来增强许多语言。对共同基准测试的实验表明,这种增强功能受益于MLLM和MLKGS,实现:(1)知识图完成的可比性或改善性能相对于基准,尤其是对于低资产阶级语言(对于知识图而言,这是不可避免的); (2)提高了需要多语言事实知识的语言理解任务的MLLM表现;同时,在其他一般语言任务上保持绩效。
Large language models appear to learn facts from the large text corpora they are trained on. Such facts are encoded implicitly within their many parameters, making it difficult to verify or manipulate what knowledge has been learned. Language models have recently been extended to multilingual language models (MLLMs), enabling knowledge to be learned across hundreds of languages. Meanwhile, knowledge graphs contain facts in an explicit triple format, which require careful and costly curation and are only available in a few high-resource languages, restricting their research and application. To address these issues, we propose to enhance MLLMs with knowledge from multilingual knowledge graphs (MLKGs) so as to tackle language and knowledge graph tasks across many languages, including low-resource ones. Specifically, we introduce a lightweight adapter set to enhance MLLMs with cross-lingual entity alignment and facts from MLKGs for many languages. Experiments on common benchmarks show that such enhancement benefits both MLLMs and MLKGs, achieving: (1) comparable or improved performance for knowledge graph completion and entity alignment relative to baselines, especially for low-resource languages (for which knowledge graphs are unavailable); and (2) improved MLLM performance on language understanding tasks that require multilingual factual knowledge; all while maintaining performance on other general language tasks.