论文标题
手电筒:有效解码器的可扩展链接预测
Flashlight: Scalable Link Prediction with Effective Decoders
论文作者
论文摘要
链接预测(LP)被认为是图形学习的重要任务,其广泛的实际应用。 LP的典型应用是检索给定源节点的最高评分邻居,例如朋友推荐。这些服务希望具有很高的推理可伸缩性,以找到低潜伏期中许多候选节点的最高评分邻居。最近有两个流行的解码器主要用于计算节点嵌入的边缘得分:HadamardMLP和DOT产品解码器。经过理论和经验分析后,我们发现HadamardMLP解码器通常对LP更有效。但是,HadamardMLP缺乏在大图上检索最高得分的邻居的可扩展性,因为据我们所知,并不存在算法来检索sublinearearightions中的HadamardMLP解码器的最高得分邻居。为了使HadamardMLP可扩展,我们建议使用手电筒算法来加速HadamardMLP的最高得分邻居检索:一种具有适应性调整的查询嵌入的近似最大内部产品搜索(MIPS)技术,该算法逐渐应用了近似最大的内部产品搜索(MIPS)技术。经验结果表明,手电筒在不牺牲有效性的情况下,在大的OGBL-Citation2数据集上将LP的推理速度提高了100倍以上。我们的工作为大规模LP应用程序铺平了道路,并通过有效的HadamArdMLP解码器大大加速了它们的推断。
Link prediction (LP) has been recognized as an important task in graph learning with its broad practical applications. A typical application of LP is to retrieve the top scoring neighbors for a given source node, such as the friend recommendation. These services desire the high inference scalability to find the top scoring neighbors from many candidate nodes at low latencies. There are two popular decoders that the recent LP models mainly use to compute the edge scores from node embeddings: the HadamardMLP and Dot Product decoders. After theoretical and empirical analysis, we find that the HadamardMLP decoders are generally more effective for LP. However, HadamardMLP lacks the scalability for retrieving top scoring neighbors on large graphs, since to the best of our knowledge, there does not exist an algorithm to retrieve the top scoring neighbors for HadamardMLP decoders in sublinear complexity. To make HadamardMLP scalable, we propose the Flashlight algorithm to accelerate the top scoring neighbor retrievals for HadamardMLP: a sublinear algorithm that progressively applies approximate maximum inner product search (MIPS) techniques with adaptively adjusted query embeddings. Empirical results show that Flashlight improves the inference speed of LP by more than 100 times on the large OGBL-CITATION2 dataset without sacrificing effectiveness. Our work paves the way for large-scale LP applications with the effective HadamardMLP decoders by greatly accelerating their inference.