Learning to Solve Integer Linear Programs with Davis-Yin Splitting

In many applications, a combinatorial problem must be repeatedly solved with similar, but distinct parameters. Yet, the parameters are not directly observed; only contextual data that correlates with is available. It is tempting to use a neural network to predict given . However, training such a model requires reconciling the discrete nature of combinatorial optimization with the gradient-based frameworks used to train neural networks. When the problem in question is an Integer Linear Program (ILP), one approach to overcome this training issue is to consider a continuous relaxation of the combinatorial problem. While existing methods utilizing this approach have shown to be highly effective on small problems, they do not always scale well to large problems. In this work, we draw on ideas from modern convex optimization to design a network and training scheme which scales effortlessly to problems with thousands of variables. Our experiments verify the computational advantage our proposed method enjoys on two representative problems, namely the shortest path problem and the knapsack problem.
View on arXiv