Scalable Knowledge Refactoring using Constrained Optimisation
- URL: http://arxiv.org/abs/2408.11530v1
- Date: Wed, 21 Aug 2024 11:12:42 GMT
- Title: Scalable Knowledge Refactoring using Constrained Optimisation
- Authors: Minghao Liu, David M. Cerna, Filipe Gouveia, Andrew Cropper,
- Abstract summary: We show that our approach can programs quicker and with more compression than the previous state-of-the-art approach, sometimes by 60%.
Our empirical results on multiple domains show that our approach can programs quicker and with more compression than the previous state-of-the-art approach, sometimes by 60%.
- Score: 18.706442683121615
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Knowledge refactoring compresses a logic program by introducing new rules. Current approaches struggle to scale to large programs. To overcome this limitation, we introduce a constrained optimisation refactoring approach. Our first key idea is to encode the problem with decision variables based on literals rather than rules. Our second key idea is to focus on linear invented rules. Our empirical results on multiple domains show that our approach can refactor programs quicker and with more compression than the previous state-of-the-art approach, sometimes by 60%.
Related papers
- Tractable Offline Learning of Regular Decision Processes [50.11277112628193]
This work studies offline Reinforcement Learning (RL) in a class of non-Markovian environments called Regular Decision Processes (RDPs)
Ins, the unknown dependency of future observations and rewards from the past interactions can be captured experimentally.
Many algorithms first reconstruct this unknown dependency using automata learning techniques.
arXiv Detail & Related papers (2024-09-04T14:26:58Z) - On Stateful Value Factorization in Multi-Agent Reinforcement Learning [19.342676562701794]
We introduce Duelmix, a factorization algorithm that learns distinct per-agent utility estimators to improve performance.
Experiments on StarCraft II micromanagement and Box Pushing tasks demonstrate the benefits of our intuitions.
arXiv Detail & Related papers (2024-08-27T19:45:26Z) - Hierarchical Context Merging: Better Long Context Understanding for Pre-trained LLMs [61.40047491337793]
We present Hierarchical cOntext MERging (HOMER), a new training-free scheme designed to overcome the limitations of large language models.
HomeR uses a divide-and-conquer algorithm, dividing long inputs into manageable chunks.
A token reduction technique precedes each merging, ensuring memory usage efficiency.
arXiv Detail & Related papers (2024-04-16T06:34:08Z) - Truly No-Regret Learning in Constrained MDPs [61.78619476991494]
We propose a model-based primal-dual algorithm to learn in an unknown CMDP.
We prove that our algorithm achieves sublinear regret without error cancellations.
arXiv Detail & Related papers (2024-02-24T09:47:46Z) - Learning logic programs by discovering higher-order abstractions [20.57989636488575]
We introduce the higher-order optimisation problem.
The goal is to compress a logic program by discovering higher-order abstractions.
We implement our approach in Stevie, which formulates the problem as a constraint problem.
arXiv Detail & Related papers (2023-08-16T12:50:10Z) - Learning to Accelerate Approximate Methods for Solving Integer
Programming via Early Fixing [29.29673962163146]
A large fraction of variables solved by some iterative approximate methods fluctuate around their final converged discrete states in very long iterations.
Inspired by this observation, we aim to accelerate these approximate methods by early fixing these fluctuated variables to their converged states.
We formulate the whole early fixing process as a Markov decision process, and train it using imitation learning.
arXiv Detail & Related papers (2022-07-05T14:46:47Z) - Learning logic programs by combining programs [24.31242130341093]
We introduce an approach where we learn small non-separable programs and combine them.
We implement our approach in a constraint-driven ILP system.
Our experiments on multiple domains, including game playing and program synthesis, show that our approach can drastically outperform existing approaches.
arXiv Detail & Related papers (2022-06-01T10:07:37Z) - Robust Predictable Control [149.71263296079388]
We show that our method achieves much tighter compression than prior methods, achieving up to 5x higher reward than a standard information bottleneck.
We also demonstrate that our method learns policies that are more robust and generalize better to new tasks.
arXiv Detail & Related papers (2021-09-07T17:29:34Z) - Data-driven Weight Initialization with Sylvester Solvers [72.11163104763071]
We propose a data-driven scheme to initialize the parameters of a deep neural network.
We show that our proposed method is especially effective in few-shot and fine-tuning settings.
arXiv Detail & Related papers (2021-05-02T07:33:16Z) - Boosting Data Reduction for the Maximum Weight Independent Set Problem
Using Increasing Transformations [59.84561168501493]
We introduce new generalized data reduction and transformation rules for the maximum weight independent set problem.
Surprisingly, these so-called increasing transformations can simplify the problem and also open up the reduction space to yield even smaller irreducible graphs later in the algorithm.
Our algorithm computes significantly smaller irreducible graphs on all except one instance, solves more instances to optimality than previously possible, is up to two orders of magnitude faster than the best state-of-the-art solver, and finds higher-quality solutions than solvers DynWVC and HILS.
arXiv Detail & Related papers (2020-08-12T08:52:50Z) - Knowledge Refactoring for Inductive Program Synthesis [37.54933305877746]
Humans constantly restructure knowledge to use it more efficiently.
Our goal is to give a machine learning system similar abilities so that it can learn more efficiently.
arXiv Detail & Related papers (2020-04-21T12:04:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.