论文标题
语言表示模型可以在赌注中思考吗?
Can Language Representation Models Think in Bets?
论文作者
论文摘要
近年来,基于变压器的语言表示模型(LRMS)已取得了最新的自然语言理解问题的结果,例如问题回答和文本摘要。由于这些模型被整合到现实世界应用中,因此评估其做出理性决策的能力是重要的研究议程,并具有实际的后果。本文通过一组精心设计的决策基准和实验研究了LRMS的合理决策能力。受认知科学的经典作品的启发,我们将决策问题对准了赌注。然后,我们研究了LRM选择具有最佳或最少阳性预期增益的结果的能力。通过对四个已建立的LRM的实验实验的强大实验,我们表明,如果模型首先在具有相同结构的BET问题上进行微调,则只能“思考”。在仍然保留其基本特征的同时,修改BET问题的结构,平均将LRM的性能降低了25 \%以上,尽管绝对性能仍然远高于随机性。当选择具有非负预期增益的结果,而不是最佳或严格的预期增益时,LRM也更为理性。我们的结果表明,LRM可能会应用于依靠认知决策技巧的任务,但是在他们可以坚固做出理性决策之前,需要进行更多的研究。
In recent years, transformer-based language representation models (LRMs) have achieved state-of-the-art results on difficult natural language understanding problems, such as question answering and text summarization. As these models are integrated into real-world applications, evaluating their ability to make rational decisions is an important research agenda, with practical ramifications. This article investigates LRMs' rational decision-making ability through a carefully designed set of decision-making benchmarks and experiments. Inspired by classic work in cognitive science, we model the decision-making problem as a bet. We then investigate an LRM's ability to choose outcomes that have optimal, or at minimum, positive expected gain. Through a robust body of experiments on four established LRMs, we show that a model is only able to `think in bets' if it is first fine-tuned on bet questions with an identical structure. Modifying the bet question's structure, while still retaining its fundamental characteristics, decreases an LRM's performance by more than 25\%, on average, although absolute performance remains well above random. LRMs are also found to be more rational when selecting outcomes with non-negative expected gain, rather than optimal or strictly positive expected gain. Our results suggest that LRMs could potentially be applied to tasks that rely on cognitive decision-making skills, but that more research is necessary before they can robustly make rational decisions.