论文标题
神经ILP体系结构中可扩展学习的无溶液框架
A Solver-Free Framework for Scalable Learning in Neural ILP Architectures
论文作者
论文摘要
最近,有一个重点是设计具有整数线性编程(ILP)层的体系结构(本文称为神经ILP)。神经ILP架构适用于需要数据驱动的约束学习或需要感知(神经)和推理(ILP)的任务的纯推理任务。最近对神经ILP端到端培训的SOTA方法明确地通过ILP黑匣子(Paulus等,2021)定义了梯度 - 由于呼吁在MiniBatch中为每个培训数据点呼叫,这训练非常缓慢。作为回应,我们提出了一种无求解器的替代培训策略,即在培训时根本不称为ILP解决方案。神经ILP具有一组可训练的超平面(用于ILP中的成本和约束),共同代表多面体。我们的关键思想是,训练损失应强加于最终的多面体将阳性(满足的所有约束)与负面(至少一个违反的约束或次优的成本值)分开,这是通过软边缘配方将其分开的。虽然作为培训数据的一部分提供了积极的例子,但我们设计了用于产生负样本的新技术。我们的解决方案足够灵活,可以处理平等以及不平等约束。与纯粹的神经基础线和其他需要基于求解的培训的最新模型相比,需要学习ILP的限制的几个问题的实验,都需要学习ILP的限制,这表明我们的方法具有出色的性能和尺度。特别是,我们能够在9 x 9的符号和视觉上获得出色的性能,而另一个神经ILP求解器无法扩展。
There is a recent focus on designing architectures that have an Integer Linear Programming (ILP) layer within a neural model (referred to as Neural ILP in this paper). Neural ILP architectures are suitable for pure reasoning tasks that require data-driven constraint learning or for tasks requiring both perception (neural) and reasoning (ILP). A recent SOTA approach for end-to-end training of Neural ILP explicitly defines gradients through the ILP black box (Paulus et al. 2021) - this trains extremely slowly, owing to a call to the underlying ILP solver for every training data point in a minibatch. In response, we present an alternative training strategy that is solver-free, i.e., does not call the ILP solver at all at training time. Neural ILP has a set of trainable hyperplanes (for cost and constraints in ILP), together representing a polyhedron. Our key idea is that the training loss should impose that the final polyhedron separates the positives (all constraints satisfied) from the negatives (at least one violated constraint or a suboptimal cost value), via a soft-margin formulation. While positive example(s) are provided as part of the training data, we devise novel techniques for generating negative samples. Our solution is flexible enough to handle equality as well as inequality constraints. Experiments on several problems, both perceptual as well as symbolic, which require learning the constraints of an ILP, show that our approach has superior performance and scales much better compared to purely neural baselines and other state-of-the-art models that require solver-based training. In particular, we are able to obtain excellent performance in 9 x 9 symbolic and visual sudoku, to which the other Neural ILP solver is not able to scale.