CombOptNet: Fit the Right NP-Hard Problem by Learning Integer
Programming Constraints
- URL: http://arxiv.org/abs/2105.02343v1
- Date: Wed, 5 May 2021 21:52:53 GMT
- Title: CombOptNet: Fit the Right NP-Hard Problem by Learning Integer
Programming Constraints
- Authors: Anselm Paulus and Michal Rol\'inek and V\'it Musil and Brandon Amos
and Georg Martius
- Abstract summary: We aim to integrate integer programming solvers into neural network architectures as layers capable of learning both the cost terms and the constraints.
The resulting end-to-end trainable architectures jointly extract features from raw data and solve a suitable (learned) problem with state-of-the-art integer programming solvers.
- Score: 20.659237363210774
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Bridging logical and algorithmic reasoning with modern machine learning
techniques is a fundamental challenge with potentially transformative impact.
On the algorithmic side, many NP-hard problems can be expressed as integer
programs, in which the constraints play the role of their "combinatorial
specification". In this work, we aim to integrate integer programming solvers
into neural network architectures as layers capable of learning both the cost
terms and the constraints. The resulting end-to-end trainable architectures
jointly extract features from raw data and solve a suitable (learned)
combinatorial problem with state-of-the-art integer programming solvers. We
demonstrate the potential of such layers with an extensive performance analysis
on synthetic data and with a demonstration on a competitive computer vision
keypoint matching benchmark.
Related papers
- Towards a Generic Representation of Combinatorial Problems for
Learning-Based Approaches [2.2526069385327316]
In recent years, there has been a growing interest in using learning-based approaches for solving problems.
The challenge lies in encoding the targeted problems into a structure compatible with the learning algorithm.
Many existing works have proposed problem-specific representations, often in the form of a graph, to leverage the advantages of textitgraph neural networks
This paper advocates for a fully generic representation of problems for learning-based approaches.
arXiv Detail & Related papers (2024-03-09T22:28:46Z) - Multi-Level GNN Preconditioner for Solving Large Scale Problems [0.0]
Graph Neural Networks (GNNs) are great for learning from unstructured data like meshes but are often limited to small-scale problems.
This paper introduces a novel preconditioner integrating a GNN model within a multi-level Domain Decomposition framework.
The proposed GNN-based preconditioner is used to enhance the efficiency of a Krylov method, resulting in a hybrid solver that can converge with any desired level of accuracy.
arXiv Detail & Related papers (2024-02-13T08:50:14Z) - When Do Program-of-Thoughts Work for Reasoning? [51.2699797837818]
We propose complexity-impacted reasoning score (CIRS) to measure correlation between code and reasoning abilities.
Specifically, we use the abstract syntax tree to encode the structural information and calculate logical complexity.
Code will be integrated into the EasyInstruct framework at https://github.com/zjunlp/EasyInstruct.
arXiv Detail & Related papers (2023-08-29T17:22:39Z) - Neural Algorithmic Reasoning for Combinatorial Optimisation [20.36694807847833]
We propose leveraging recent advancements in neural reasoning to improve the learning of CO problems.
We suggest pre-training our neural model on relevant algorithms before training it on CO instances.
Our results demonstrate that by using this learning setup, we achieve superior performance compared to non-algorithmically informed deep learning models.
arXiv Detail & Related papers (2023-05-18T13:59:02Z) - Scalable Coupling of Deep Learning with Logical Reasoning [0.0]
We introduce a scalable neural architecture and loss function dedicated to learning the constraints and criteria of NP-hard reasoning problems.
Our loss function solves one of the main limitations of Besag's pseudo-loglikelihood, enabling learning of high energies.
arXiv Detail & Related papers (2023-05-12T17:09:34Z) - Tunable Complexity Benchmarks for Evaluating Physics-Informed Neural
Networks on Coupled Ordinary Differential Equations [64.78260098263489]
In this work, we assess the ability of physics-informed neural networks (PINNs) to solve increasingly-complex coupled ordinary differential equations (ODEs)
We show that PINNs eventually fail to produce correct solutions to these benchmarks as their complexity increases.
We identify several reasons why this may be the case, including insufficient network capacity, poor conditioning of the ODEs, and high local curvature, as measured by the Laplacian of the PINN loss.
arXiv Detail & Related papers (2022-10-14T15:01:32Z) - Contextual Model Aggregation for Fast and Robust Federated Learning in
Edge Computing [88.76112371510999]
Federated learning is a prime candidate for distributed machine learning at the network edge.
Existing algorithms face issues with slow convergence and/or robustness of performance.
We propose a contextual aggregation scheme that achieves the optimal context-dependent bound on loss reduction.
arXiv Detail & Related papers (2022-03-23T21:42:31Z) - A Differentiable Approach to Combinatorial Optimization using Dataless
Neural Networks [20.170140039052455]
We propose a radically different approach in that no data is required for training the neural networks that produce the solution.
In particular, we reduce the optimization problem to a neural network and employ a dataless training scheme to refine the parameters of the network such that those parameters yield the structure of interest.
arXiv Detail & Related papers (2022-03-15T19:21:31Z) - Strong Generalization and Efficiency in Neural Programs [69.18742158883869]
We study the problem of learning efficient algorithms that strongly generalize in the framework of neural program induction.
By carefully designing the input / output interfaces of the neural model and through imitation, we are able to learn models that produce correct results for arbitrary input sizes.
arXiv Detail & Related papers (2020-07-07T17:03:02Z) - An Integer Linear Programming Framework for Mining Constraints from Data [81.60135973848125]
We present a general framework for mining constraints from data.
In particular, we consider the inference in structured output prediction as an integer linear programming (ILP) problem.
We show that our approach can learn to solve 9x9 Sudoku puzzles and minimal spanning tree problems from examples without providing the underlying rules.
arXiv Detail & Related papers (2020-06-18T20:09:53Z) - Machine Number Sense: A Dataset of Visual Arithmetic Problems for
Abstract and Relational Reasoning [95.18337034090648]
We propose a dataset, Machine Number Sense (MNS), consisting of visual arithmetic problems automatically generated using a grammar model--And-Or Graph (AOG)
These visual arithmetic problems are in the form of geometric figures.
We benchmark the MNS dataset using four predominant neural network models as baselines in this visual reasoning task.
arXiv Detail & Related papers (2020-04-25T17:14:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.