论文标题
学习稀疏图形均值野外游戏
Learning Sparse Graphon Mean Field Games
论文作者
论文摘要
尽管在过去几年中,多机构增强学习(MARL)的领域取得了长足的进步,但解决了许多代理的系统仍然是一个艰巨的挑战。 Graphon平均场游戏(GMFGS)可实现对MARL问题的可扩展分析,而MARL问题原本是棘手的。通过图形的数学结构,这种方法仅限于密集的图形,这些图形不足以描述许多现实世界网络,例如幂律图。我们的论文介绍了GMFGS的新型公式,称为LPGMFGS,该公式利用了$ l^p $图形的图理论概念,并提供了一种机器学习工具,以有效,准确地近似于稀疏网络问题的解决方案。这尤其包括在各个应用领域经验观察到的电力法网络,并且不能由标准图形捕获。我们得出理论的存在和融合保证,并提供了经验示例,以证明我们与许多代理的系统学习方法的准确性。此外,我们将在线镜下降(OMD)学习算法扩展到我们的设置,以加速学习速度,经验显示其能力,并使用平滑的步骤图形的新颖概念进行理论分析。通常,我们为在众多研究领域提供了一大批原本棘手的问题,提供了一种可扩展的,数学上有充分的机器学习方法。
Although the field of multi-agent reinforcement learning (MARL) has made considerable progress in the last years, solving systems with a large number of agents remains a hard challenge. Graphon mean field games (GMFGs) enable the scalable analysis of MARL problems that are otherwise intractable. By the mathematical structure of graphons, this approach is limited to dense graphs which are insufficient to describe many real-world networks such as power law graphs. Our paper introduces a novel formulation of GMFGs, called LPGMFGs, which leverages the graph theoretical concept of $L^p$ graphons and provides a machine learning tool to efficiently and accurately approximate solutions for sparse network problems. This especially includes power law networks which are empirically observed in various application areas and cannot be captured by standard graphons. We derive theoretical existence and convergence guarantees and give empirical examples that demonstrate the accuracy of our learning approach for systems with many agents. Furthermore, we extend the Online Mirror Descent (OMD) learning algorithm to our setup to accelerate learning speed, empirically show its capabilities, and conduct a theoretical analysis using the novel concept of smoothed step graphons. In general, we provide a scalable, mathematically well-founded machine learning approach to a large class of otherwise intractable problems of great relevance in numerous research fields.